var/home/core/zuul-output/0000755000175000017500000000000015114725423014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114733242015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003653240215114733233017704 0ustar rootrootDec 06 04:06:50 crc systemd[1]: Starting Kubernetes Kubelet... Dec 06 04:06:50 crc restorecon[4706]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 04:06:50 crc restorecon[4706]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 06 04:06:51 crc kubenswrapper[4718]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 04:06:51 crc kubenswrapper[4718]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 06 04:06:51 crc kubenswrapper[4718]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 04:06:51 crc kubenswrapper[4718]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 04:06:51 crc kubenswrapper[4718]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 06 04:06:51 crc kubenswrapper[4718]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.131601 4718 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138710 4718 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138750 4718 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138761 4718 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138771 4718 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138782 4718 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138793 4718 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138804 4718 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138814 4718 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138824 4718 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138833 4718 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138842 4718 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138850 4718 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138858 4718 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138866 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138874 4718 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138882 4718 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138889 4718 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138897 4718 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138905 4718 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138913 4718 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138920 4718 feature_gate.go:330] unrecognized feature gate: Example Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138928 4718 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138936 4718 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138944 4718 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138962 4718 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138970 4718 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138978 4718 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138986 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.138994 4718 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139002 4718 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139010 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139017 4718 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139026 4718 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139036 4718 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139044 4718 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139055 4718 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139065 4718 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139074 4718 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139083 4718 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139092 4718 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139100 4718 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139107 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139115 4718 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139122 4718 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139166 4718 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139175 4718 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139182 4718 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139190 4718 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139197 4718 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139204 4718 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139212 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139220 4718 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139255 4718 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139264 4718 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139272 4718 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139280 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139288 4718 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139299 4718 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139307 4718 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139315 4718 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139323 4718 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139331 4718 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139339 4718 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139346 4718 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139354 4718 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139367 4718 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139377 4718 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139386 4718 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139393 4718 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139402 4718 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.139409 4718 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139879 4718 flags.go:64] FLAG: --address="0.0.0.0" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139906 4718 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139922 4718 flags.go:64] FLAG: --anonymous-auth="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139935 4718 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139946 4718 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139956 4718 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139968 4718 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139979 4718 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139988 4718 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.139997 4718 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140007 4718 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140019 4718 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140028 4718 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140037 4718 flags.go:64] FLAG: --cgroup-root="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140046 4718 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140055 4718 flags.go:64] FLAG: --client-ca-file="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140064 4718 flags.go:64] FLAG: --cloud-config="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140072 4718 flags.go:64] FLAG: --cloud-provider="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140081 4718 flags.go:64] FLAG: --cluster-dns="[]" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140092 4718 flags.go:64] FLAG: --cluster-domain="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140101 4718 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140111 4718 flags.go:64] FLAG: --config-dir="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140119 4718 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140129 4718 flags.go:64] FLAG: --container-log-max-files="5" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140140 4718 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140149 4718 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140158 4718 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140168 4718 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140177 4718 flags.go:64] FLAG: --contention-profiling="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140186 4718 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140195 4718 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140204 4718 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140213 4718 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140225 4718 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140270 4718 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140279 4718 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140288 4718 flags.go:64] FLAG: --enable-load-reader="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140297 4718 flags.go:64] FLAG: --enable-server="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140306 4718 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140318 4718 flags.go:64] FLAG: --event-burst="100" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140327 4718 flags.go:64] FLAG: --event-qps="50" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140336 4718 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140345 4718 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140354 4718 flags.go:64] FLAG: --eviction-hard="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140365 4718 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140374 4718 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140383 4718 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140394 4718 flags.go:64] FLAG: --eviction-soft="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140405 4718 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140413 4718 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140422 4718 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140431 4718 flags.go:64] FLAG: --experimental-mounter-path="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140440 4718 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140449 4718 flags.go:64] FLAG: --fail-swap-on="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140457 4718 flags.go:64] FLAG: --feature-gates="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140468 4718 flags.go:64] FLAG: --file-check-frequency="20s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140477 4718 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140487 4718 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140498 4718 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140507 4718 flags.go:64] FLAG: --healthz-port="10248" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140516 4718 flags.go:64] FLAG: --help="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140525 4718 flags.go:64] FLAG: --hostname-override="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140533 4718 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140543 4718 flags.go:64] FLAG: --http-check-frequency="20s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140552 4718 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140560 4718 flags.go:64] FLAG: --image-credential-provider-config="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140569 4718 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140578 4718 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140587 4718 flags.go:64] FLAG: --image-service-endpoint="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140596 4718 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140605 4718 flags.go:64] FLAG: --kube-api-burst="100" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140613 4718 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140623 4718 flags.go:64] FLAG: --kube-api-qps="50" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140632 4718 flags.go:64] FLAG: --kube-reserved="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140641 4718 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140650 4718 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140659 4718 flags.go:64] FLAG: --kubelet-cgroups="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140668 4718 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140676 4718 flags.go:64] FLAG: --lock-file="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140685 4718 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140694 4718 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140703 4718 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140727 4718 flags.go:64] FLAG: --log-json-split-stream="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140737 4718 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140746 4718 flags.go:64] FLAG: --log-text-split-stream="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140755 4718 flags.go:64] FLAG: --logging-format="text" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140764 4718 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140773 4718 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140782 4718 flags.go:64] FLAG: --manifest-url="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140791 4718 flags.go:64] FLAG: --manifest-url-header="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140803 4718 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140812 4718 flags.go:64] FLAG: --max-open-files="1000000" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140823 4718 flags.go:64] FLAG: --max-pods="110" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140832 4718 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140843 4718 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140854 4718 flags.go:64] FLAG: --memory-manager-policy="None" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140865 4718 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140876 4718 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140887 4718 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140899 4718 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140925 4718 flags.go:64] FLAG: --node-status-max-images="50" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140936 4718 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140945 4718 flags.go:64] FLAG: --oom-score-adj="-999" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140955 4718 flags.go:64] FLAG: --pod-cidr="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140963 4718 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140984 4718 flags.go:64] FLAG: --pod-manifest-path="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.140993 4718 flags.go:64] FLAG: --pod-max-pids="-1" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141001 4718 flags.go:64] FLAG: --pods-per-core="0" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141011 4718 flags.go:64] FLAG: --port="10250" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141020 4718 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141029 4718 flags.go:64] FLAG: --provider-id="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141038 4718 flags.go:64] FLAG: --qos-reserved="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141046 4718 flags.go:64] FLAG: --read-only-port="10255" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141055 4718 flags.go:64] FLAG: --register-node="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141064 4718 flags.go:64] FLAG: --register-schedulable="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141074 4718 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141089 4718 flags.go:64] FLAG: --registry-burst="10" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141098 4718 flags.go:64] FLAG: --registry-qps="5" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141107 4718 flags.go:64] FLAG: --reserved-cpus="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141117 4718 flags.go:64] FLAG: --reserved-memory="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141129 4718 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141138 4718 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141147 4718 flags.go:64] FLAG: --rotate-certificates="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141155 4718 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141164 4718 flags.go:64] FLAG: --runonce="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141173 4718 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141182 4718 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141191 4718 flags.go:64] FLAG: --seccomp-default="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141200 4718 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141209 4718 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141218 4718 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141253 4718 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141263 4718 flags.go:64] FLAG: --storage-driver-password="root" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141272 4718 flags.go:64] FLAG: --storage-driver-secure="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141281 4718 flags.go:64] FLAG: --storage-driver-table="stats" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141290 4718 flags.go:64] FLAG: --storage-driver-user="root" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141299 4718 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141309 4718 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141318 4718 flags.go:64] FLAG: --system-cgroups="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141327 4718 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141340 4718 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141350 4718 flags.go:64] FLAG: --tls-cert-file="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141358 4718 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141370 4718 flags.go:64] FLAG: --tls-min-version="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141379 4718 flags.go:64] FLAG: --tls-private-key-file="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141388 4718 flags.go:64] FLAG: --topology-manager-policy="none" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141397 4718 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141410 4718 flags.go:64] FLAG: --topology-manager-scope="container" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141419 4718 flags.go:64] FLAG: --v="2" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141431 4718 flags.go:64] FLAG: --version="false" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141442 4718 flags.go:64] FLAG: --vmodule="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141453 4718 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.141462 4718 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141705 4718 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141715 4718 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141725 4718 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141733 4718 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141741 4718 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141750 4718 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141757 4718 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141765 4718 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141773 4718 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141781 4718 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141788 4718 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141796 4718 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141804 4718 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141812 4718 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141820 4718 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141827 4718 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141835 4718 feature_gate.go:330] unrecognized feature gate: Example Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141842 4718 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141853 4718 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141864 4718 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141873 4718 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141882 4718 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141891 4718 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141899 4718 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141908 4718 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141916 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141927 4718 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141936 4718 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141944 4718 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141952 4718 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141961 4718 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141969 4718 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141977 4718 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141984 4718 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.141992 4718 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142000 4718 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142008 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142015 4718 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142024 4718 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142031 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142039 4718 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142047 4718 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142054 4718 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142062 4718 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142069 4718 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142077 4718 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142085 4718 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142092 4718 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142100 4718 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142108 4718 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142116 4718 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142123 4718 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142131 4718 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142138 4718 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142146 4718 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142153 4718 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142161 4718 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142168 4718 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142181 4718 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142191 4718 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142200 4718 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142211 4718 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142221 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142260 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142272 4718 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142282 4718 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142293 4718 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142303 4718 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142312 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142323 4718 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.142332 4718 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.142360 4718 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.155068 4718 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.155106 4718 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155376 4718 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155397 4718 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155409 4718 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155420 4718 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155432 4718 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155443 4718 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155454 4718 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155465 4718 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155476 4718 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155487 4718 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155496 4718 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155507 4718 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155517 4718 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155528 4718 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155538 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155548 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155558 4718 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155572 4718 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155586 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155597 4718 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155608 4718 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155619 4718 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155630 4718 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155646 4718 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155658 4718 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155670 4718 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155681 4718 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155692 4718 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155703 4718 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155713 4718 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155723 4718 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155733 4718 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155744 4718 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155756 4718 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155771 4718 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155781 4718 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155791 4718 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155801 4718 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155811 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155821 4718 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155831 4718 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155841 4718 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155851 4718 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155861 4718 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155871 4718 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155882 4718 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155892 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155902 4718 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155913 4718 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155924 4718 feature_gate.go:330] unrecognized feature gate: Example Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155937 4718 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155951 4718 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155962 4718 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155973 4718 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155984 4718 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.155995 4718 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156005 4718 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156015 4718 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156026 4718 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156036 4718 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156046 4718 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156056 4718 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156067 4718 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156078 4718 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156089 4718 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156099 4718 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156108 4718 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156118 4718 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156129 4718 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156140 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156169 4718 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.156186 4718 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156569 4718 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156595 4718 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156608 4718 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156618 4718 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156648 4718 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156663 4718 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156674 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156687 4718 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156701 4718 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156712 4718 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156722 4718 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156732 4718 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156743 4718 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156752 4718 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156762 4718 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156772 4718 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156782 4718 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156792 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156802 4718 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156812 4718 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156823 4718 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156833 4718 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156844 4718 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156854 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156864 4718 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156874 4718 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156883 4718 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156893 4718 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156903 4718 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156914 4718 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156924 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156934 4718 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156944 4718 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156954 4718 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156981 4718 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.156992 4718 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157002 4718 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157012 4718 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157023 4718 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157033 4718 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157043 4718 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157053 4718 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157066 4718 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157080 4718 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157092 4718 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157103 4718 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157113 4718 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157123 4718 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157133 4718 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157144 4718 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157156 4718 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157166 4718 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157176 4718 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157186 4718 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157196 4718 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157206 4718 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157216 4718 feature_gate.go:330] unrecognized feature gate: Example Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157226 4718 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157271 4718 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157282 4718 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157291 4718 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157305 4718 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157318 4718 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157329 4718 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157341 4718 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157352 4718 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157362 4718 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157372 4718 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157382 4718 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157392 4718 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.157405 4718 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.157421 4718 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.157813 4718 server.go:940] "Client rotation is on, will bootstrap in background" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.164211 4718 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.164398 4718 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.165340 4718 server.go:997] "Starting client certificate rotation" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.165388 4718 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.165650 4718 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-18 09:08:26.883792112 +0000 UTC Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.165807 4718 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 293h1m35.717993801s for next certificate rotation Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.174524 4718 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.177615 4718 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.186023 4718 log.go:25] "Validated CRI v1 runtime API" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.209020 4718 log.go:25] "Validated CRI v1 image API" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.210642 4718 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.214312 4718 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-06-04-02-34-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.214357 4718 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.243354 4718 manager.go:217] Machine: {Timestamp:2025-12-06 04:06:51.241272102 +0000 UTC m=+0.246977333 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:76af1561-f0fa-462c-8fb7-968628565864 BootID:3f36729d-94a7-4ff7-b13e-5910bb9e8ea4 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:71:16:b5 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:71:16:b5 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:72:98:f5 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:3f:d4:81 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:3e:a3:98 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:5b:bf:1a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d2:a9:b6:50:99:11 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:66:76:d7:a5:e3:05 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.243769 4718 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.244037 4718 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.244795 4718 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.245084 4718 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.245138 4718 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.245491 4718 topology_manager.go:138] "Creating topology manager with none policy" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.245511 4718 container_manager_linux.go:303] "Creating device plugin manager" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.245829 4718 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.245894 4718 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.246283 4718 state_mem.go:36] "Initialized new in-memory state store" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.246423 4718 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.247394 4718 kubelet.go:418] "Attempting to sync node with API server" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.247427 4718 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.247463 4718 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.247482 4718 kubelet.go:324] "Adding apiserver pod source" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.247577 4718 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.249906 4718 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.250436 4718 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.251766 4718 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.251859 4718 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.251961 4718 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.251965 4718 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252088 4718 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252818 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252849 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252858 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252867 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252883 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252893 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252903 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252919 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252933 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252946 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252963 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.252974 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.253255 4718 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.253895 4718 server.go:1280] "Started kubelet" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.254525 4718 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.254925 4718 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.254864 4718 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:51 crc systemd[1]: Started Kubernetes Kubelet. Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.257153 4718 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.257195 4718 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.257517 4718 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 03:40:31.049227197 +0000 UTC Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.257591 4718 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 311h33m39.791642246s for next certificate rotation Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.257706 4718 server.go:460] "Adding debug handlers to kubelet server" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.266488 4718 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.266694 4718 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.266702 4718 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.266982 4718 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.267665 4718 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.267748 4718 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.267930 4718 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.268992 4718 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e84b3af7a4859 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 04:06:51.253852249 +0000 UTC m=+0.259557420,LastTimestamp:2025-12-06 04:06:51.253852249 +0000 UTC m=+0.259557420,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.270221 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="200ms" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.273345 4718 factory.go:153] Registering CRI-O factory Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.273381 4718 factory.go:221] Registration of the crio container factory successfully Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.273469 4718 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.273482 4718 factory.go:55] Registering systemd factory Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.273489 4718 factory.go:221] Registration of the systemd container factory successfully Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.273517 4718 factory.go:103] Registering Raw factory Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.273532 4718 manager.go:1196] Started watching for new ooms in manager Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.274262 4718 manager.go:319] Starting recovery of all containers Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.287885 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.287963 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.287987 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.288007 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.288026 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.288052 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289385 4718 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289448 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289479 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289515 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289543 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289570 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289596 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289628 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289659 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289687 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289715 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289742 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289768 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289846 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289871 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289896 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289921 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289946 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289971 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.289995 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290068 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290107 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290139 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290171 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290197 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290224 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290337 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290366 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290394 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290421 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290449 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290476 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290503 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290530 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290559 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290586 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290629 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290664 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290691 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290720 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290747 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290776 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290807 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290837 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290865 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290893 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290920 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290956 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.290990 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291022 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291054 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291085 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291159 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291192 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291217 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291282 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291312 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291342 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291373 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291403 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291430 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291457 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291483 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291511 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291540 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291570 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291599 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291628 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291656 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291684 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291713 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291746 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291763 4718 manager.go:324] Recovery completed Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.291774 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.292713 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.292759 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.292792 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.292822 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.292855 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.292883 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.292910 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.292939 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.292968 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293001 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293026 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293054 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293085 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293112 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293138 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293165 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293190 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293216 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293282 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293311 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293338 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293368 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293397 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293426 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293457 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293488 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293527 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293559 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293596 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293633 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293665 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293700 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293732 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.293762 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295367 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295464 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295510 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295561 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295593 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295630 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295684 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295714 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295760 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295793 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295826 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295869 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295903 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295949 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.295980 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296009 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296051 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296083 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296113 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296153 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296180 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296223 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296288 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296318 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296364 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296391 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296432 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296461 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296494 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296584 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296619 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296662 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296696 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.296729 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297201 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297265 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297309 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297343 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297374 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297411 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297439 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297466 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297503 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297536 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297581 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297610 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297636 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297677 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297707 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297746 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297774 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297801 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297838 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297866 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297902 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297930 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.297961 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298000 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298030 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298068 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298096 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298124 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298166 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298195 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298270 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298306 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298362 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298400 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298431 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298461 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298502 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298528 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298566 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298593 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298623 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298663 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298689 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298728 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298755 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298784 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298825 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298857 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298894 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298922 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298948 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.298988 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299015 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299055 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299092 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299156 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299201 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299261 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299303 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299329 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299359 4718 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299386 4718 reconstruct.go:97] "Volume reconstruction finished" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.299405 4718 reconciler.go:26] "Reconciler: start to sync state" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.315389 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.317866 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.317902 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.317914 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.319066 4718 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.319082 4718 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.319103 4718 state_mem.go:36] "Initialized new in-memory state store" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.325040 4718 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.326838 4718 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.326938 4718 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.326979 4718 kubelet.go:2335] "Starting kubelet main sync loop" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.327082 4718 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.328454 4718 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.328662 4718 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.328547 4718 policy_none.go:49] "None policy: Start" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.331223 4718 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.331269 4718 state_mem.go:35] "Initializing new in-memory state store" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.366569 4718 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.385465 4718 manager.go:334] "Starting Device Plugin manager" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.385577 4718 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.385597 4718 server.go:79] "Starting device plugin registration server" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.386178 4718 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.386203 4718 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.387160 4718 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.387336 4718 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.387353 4718 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.392917 4718 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.428073 4718 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.428190 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.429299 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.429354 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.429372 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.429572 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.429759 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.429806 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.430603 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.430636 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.430647 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.430766 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.430836 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.430878 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.430890 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.430917 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.430953 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.431583 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.431625 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.431638 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.431749 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.431774 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.431785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.431802 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.432024 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.432165 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.433112 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.433166 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.433180 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.433338 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.433353 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.433372 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.433383 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.433503 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.433561 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.434206 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.434252 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.434264 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.434403 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.434430 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.435440 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.435475 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.435486 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.435494 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.435511 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.435527 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.471415 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="400ms" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.487596 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.488860 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.488938 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.488950 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.488989 4718 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.489767 4718 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.501993 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502040 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502074 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502106 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502138 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502170 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502201 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502286 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502334 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502393 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502434 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502494 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502542 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502584 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.502611 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603539 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603621 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603666 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603702 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603737 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603771 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603803 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603836 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603872 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603905 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603885 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603978 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603983 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603989 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604053 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603941 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603991 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604044 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603911 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603937 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603965 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.603928 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604283 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604322 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604355 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604369 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604385 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604396 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604429 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.604550 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.690609 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.692926 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.692980 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.692998 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.693038 4718 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.693662 4718 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.759200 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.778192 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.800046 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.811473 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: I1206 04:06:51.817567 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.818728 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-1c7d62e62c1699ea9126a37d38627b09233437d3fe4e858c517c31536829fd5f WatchSource:0}: Error finding container 1c7d62e62c1699ea9126a37d38627b09233437d3fe4e858c517c31536829fd5f: Status 404 returned error can't find the container with id 1c7d62e62c1699ea9126a37d38627b09233437d3fe4e858c517c31536829fd5f Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.829918 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-7df53538e6f0eebbfd2d49d62641de0542f5f1b4aa1c63af331b479d937300e3 WatchSource:0}: Error finding container 7df53538e6f0eebbfd2d49d62641de0542f5f1b4aa1c63af331b479d937300e3: Status 404 returned error can't find the container with id 7df53538e6f0eebbfd2d49d62641de0542f5f1b4aa1c63af331b479d937300e3 Dec 06 04:06:51 crc kubenswrapper[4718]: W1206 04:06:51.849684 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ed49bf1ee98d38594cb14e90a463769c433e1ca5b9446b526b21ca45275ae04d WatchSource:0}: Error finding container ed49bf1ee98d38594cb14e90a463769c433e1ca5b9446b526b21ca45275ae04d: Status 404 returned error can't find the container with id ed49bf1ee98d38594cb14e90a463769c433e1ca5b9446b526b21ca45275ae04d Dec 06 04:06:51 crc kubenswrapper[4718]: E1206 04:06:51.872218 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="800ms" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.093956 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.095129 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.095163 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.095172 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.095196 4718 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 04:06:52 crc kubenswrapper[4718]: E1206 04:06:52.095583 4718 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Dec 06 04:06:52 crc kubenswrapper[4718]: W1206 04:06:52.223672 4718 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:52 crc kubenswrapper[4718]: E1206 04:06:52.223801 4718 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Dec 06 04:06:52 crc kubenswrapper[4718]: W1206 04:06:52.224960 4718 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:52 crc kubenswrapper[4718]: E1206 04:06:52.225020 4718 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.257326 4718 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.336600 4718 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828" exitCode=0 Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.336718 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.336904 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"dac934bf5320f0750ea0977cf169decc700a665ce9bd6d079211932e9767e263"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.337060 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.339072 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.339139 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.339152 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.340822 4718 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7" exitCode=0 Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.340922 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.340964 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b608432f1b6addf4f12116fb6b95410a8ff10822a54f3ef7833b85a4546a7b3b"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.341120 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.342996 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.343074 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.343102 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.347207 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.348013 4718 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13" exitCode=0 Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.348287 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ed49bf1ee98d38594cb14e90a463769c433e1ca5b9446b526b21ca45275ae04d"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.348576 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.350127 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.350159 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.350178 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.350429 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.350472 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7df53538e6f0eebbfd2d49d62641de0542f5f1b4aa1c63af331b479d937300e3"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.351973 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.353281 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.353340 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.353361 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.353849 4718 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2f93a418fdbe29b396f5ab029186f73885d9ce17644213dcbdc78cd79fad26b8" exitCode=0 Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.353925 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2f93a418fdbe29b396f5ab029186f73885d9ce17644213dcbdc78cd79fad26b8"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.353980 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1c7d62e62c1699ea9126a37d38627b09233437d3fe4e858c517c31536829fd5f"} Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.354151 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.355571 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.355619 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.355632 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:52 crc kubenswrapper[4718]: W1206 04:06:52.464378 4718 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:52 crc kubenswrapper[4718]: E1206 04:06:52.464532 4718 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Dec 06 04:06:52 crc kubenswrapper[4718]: W1206 04:06:52.484044 4718 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Dec 06 04:06:52 crc kubenswrapper[4718]: E1206 04:06:52.484155 4718 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Dec 06 04:06:52 crc kubenswrapper[4718]: E1206 04:06:52.674186 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="1.6s" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.895681 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.898362 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.898425 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.898435 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:52 crc kubenswrapper[4718]: I1206 04:06:52.898471 4718 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 04:06:52 crc kubenswrapper[4718]: E1206 04:06:52.900332 4718 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.359856 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.359923 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.359935 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.359946 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.363072 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.363115 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.363125 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.363132 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.363952 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.363980 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.363989 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.364797 4718 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f24e63491dde1c5b77b1d38a34b0cebb2ba3533182085243c3124ac10172211d" exitCode=0 Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.364849 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f24e63491dde1c5b77b1d38a34b0cebb2ba3533182085243c3124ac10172211d"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.364944 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.365775 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.365804 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.365814 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.367278 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5dee1caa09708aba1318dc9c5f58cae75a93d70cbeb302d6b72b62d3842a1066"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.367352 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.368167 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.368195 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.368203 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.369625 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.369680 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.369700 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d"} Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.369801 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.370601 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.370617 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.370624 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:53 crc kubenswrapper[4718]: I1206 04:06:53.914614 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.377762 4718 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ce33a6d5b891f0e72bc21fc4a67a666fba49065aa25557f394da11d32f9f23c3" exitCode=0 Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.377862 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ce33a6d5b891f0e72bc21fc4a67a666fba49065aa25557f394da11d32f9f23c3"} Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.378031 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.379410 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.379475 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.379514 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.383067 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8"} Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.383114 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.383184 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.384486 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.384512 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.384526 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.384575 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.384601 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.384613 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.500488 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.502026 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.502081 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.502201 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:54 crc kubenswrapper[4718]: I1206 04:06:54.502268 4718 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.390295 4718 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.390288 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"44bbac8cdb37f782e0571973fe03ca0ee9107ded5b1e33af09743f100f4604db"} Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.390342 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.390363 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5d935d8181e9c5e8919ea10773c2ded887fafca262dcf34753c1297d0528e2a6"} Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.390390 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"596ad2118455a55ebd93f229515a2497846beb8e33350ed20f6a0c786452e283"} Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.390404 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.391282 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.391329 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.391346 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.391467 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.391531 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:55 crc kubenswrapper[4718]: I1206 04:06:55.391555 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.397798 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"143902508bbd171b109ae7ac78a40f7247f6d540b33dda546bc9cc714794519f"} Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.397892 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"520e1368cccf5aaa459f155f5298bbec4f94ba8a349cb399b0451df691fae141"} Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.398115 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.399883 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.399943 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.399967 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.837654 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.837890 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.840311 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.840364 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.840378 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:56 crc kubenswrapper[4718]: I1206 04:06:56.844765 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.400434 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.400481 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.401706 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.401737 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.401735 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.401766 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.401778 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.401745 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.496011 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.547306 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.547467 4718 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.547504 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.548722 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.548756 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:57 crc kubenswrapper[4718]: I1206 04:06:57.548764 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:58 crc kubenswrapper[4718]: I1206 04:06:58.403457 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:58 crc kubenswrapper[4718]: I1206 04:06:58.404781 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:58 crc kubenswrapper[4718]: I1206 04:06:58.404832 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:58 crc kubenswrapper[4718]: I1206 04:06:58.404844 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.013960 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.014149 4718 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.014205 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.015927 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.015995 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.016018 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.527759 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.528035 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.529757 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.529815 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:06:59 crc kubenswrapper[4718]: I1206 04:06:59.529834 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:00 crc kubenswrapper[4718]: I1206 04:07:00.843163 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:07:00 crc kubenswrapper[4718]: I1206 04:07:00.843555 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:07:00 crc kubenswrapper[4718]: I1206 04:07:00.845469 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:00 crc kubenswrapper[4718]: I1206 04:07:00.845565 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:00 crc kubenswrapper[4718]: I1206 04:07:00.845588 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:01 crc kubenswrapper[4718]: E1206 04:07:01.393055 4718 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 04:07:01 crc kubenswrapper[4718]: I1206 04:07:01.558358 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:07:01 crc kubenswrapper[4718]: I1206 04:07:01.558614 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:07:01 crc kubenswrapper[4718]: I1206 04:07:01.560205 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:01 crc kubenswrapper[4718]: I1206 04:07:01.560295 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:01 crc kubenswrapper[4718]: I1206 04:07:01.560313 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.095744 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.095991 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.097718 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.097785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.097810 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.101615 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.257420 4718 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.417419 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.419404 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.419475 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.419497 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:03 crc kubenswrapper[4718]: W1206 04:07:03.849619 4718 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 06 04:07:03 crc kubenswrapper[4718]: I1206 04:07:03.849709 4718 trace.go:236] Trace[178825535]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 04:06:53.848) (total time: 10001ms): Dec 06 04:07:03 crc kubenswrapper[4718]: Trace[178825535]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (04:07:03.849) Dec 06 04:07:03 crc kubenswrapper[4718]: Trace[178825535]: [10.001328652s] [10.001328652s] END Dec 06 04:07:03 crc kubenswrapper[4718]: E1206 04:07:03.849730 4718 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.085738 4718 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.085799 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.118789 4718 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.118880 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.629929 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.630224 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.631948 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.632043 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.632075 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:04 crc kubenswrapper[4718]: I1206 04:07:04.663565 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 06 04:07:05 crc kubenswrapper[4718]: I1206 04:07:05.422028 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:07:05 crc kubenswrapper[4718]: I1206 04:07:05.423925 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:05 crc kubenswrapper[4718]: I1206 04:07:05.423972 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:05 crc kubenswrapper[4718]: I1206 04:07:05.423985 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:05 crc kubenswrapper[4718]: I1206 04:07:05.437775 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 06 04:07:06 crc kubenswrapper[4718]: I1206 04:07:06.096790 4718 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 04:07:06 crc kubenswrapper[4718]: I1206 04:07:06.097289 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 04:07:06 crc kubenswrapper[4718]: I1206 04:07:06.424107 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:07:06 crc kubenswrapper[4718]: I1206 04:07:06.425402 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:06 crc kubenswrapper[4718]: I1206 04:07:06.425538 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:06 crc kubenswrapper[4718]: I1206 04:07:06.425636 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.020008 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.020335 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.022004 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.022078 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.022097 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.028288 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.088503 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.090168 4718 trace.go:236] Trace[1117431930]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 04:06:54.288) (total time: 14801ms): Dec 06 04:07:09 crc kubenswrapper[4718]: Trace[1117431930]: ---"Objects listed" error: 14801ms (04:07:09.090) Dec 06 04:07:09 crc kubenswrapper[4718]: Trace[1117431930]: [14.801470757s] [14.801470757s] END Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.090185 4718 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.090524 4718 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.090666 4718 trace.go:236] Trace[1876200878]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 04:06:54.859) (total time: 14230ms): Dec 06 04:07:09 crc kubenswrapper[4718]: Trace[1876200878]: ---"Objects listed" error: 14230ms (04:07:09.090) Dec 06 04:07:09 crc kubenswrapper[4718]: Trace[1876200878]: [14.230594717s] [14.230594717s] END Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.090709 4718 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.091937 4718 trace.go:236] Trace[1137961151]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 04:06:55.517) (total time: 13573ms): Dec 06 04:07:09 crc kubenswrapper[4718]: Trace[1137961151]: ---"Objects listed" error: 13573ms (04:07:09.091) Dec 06 04:07:09 crc kubenswrapper[4718]: Trace[1137961151]: [13.573836814s] [13.573836814s] END Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.091979 4718 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.093375 4718 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.257992 4718 apiserver.go:52] "Watching apiserver" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.261980 4718 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.262432 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.262963 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.263307 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.263066 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.263406 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.263629 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.263842 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.264025 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.264136 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.264191 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.266973 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.267110 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.267186 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.266983 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.267530 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.267976 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.268150 4718 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.267999 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.268051 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.271954 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.291682 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.291780 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.291831 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.291882 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.291934 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.291991 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292042 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292090 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292105 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292141 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292209 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292302 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292395 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292441 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292490 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292578 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292635 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292719 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.292971 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293028 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293078 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293126 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293184 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293287 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293314 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293342 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293396 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293448 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293502 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293526 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293585 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293640 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293696 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293758 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293774 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.293810 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294020 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294039 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294155 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294166 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294191 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294245 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294275 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294305 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294335 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294365 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294420 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294450 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294481 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294511 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294541 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294572 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294601 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294635 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294670 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294696 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294728 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294756 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294788 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294817 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294848 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294878 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294911 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294940 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294971 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295001 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295029 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295060 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295089 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295118 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295149 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295216 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295324 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295358 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295388 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295415 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295445 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295475 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295504 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295533 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295593 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295624 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295652 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295682 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295712 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295741 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295770 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295798 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295829 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295855 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295885 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295916 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295947 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295977 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296010 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296040 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296070 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296100 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296128 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296160 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296190 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296218 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296273 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296335 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296365 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296395 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296427 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296458 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296489 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296520 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296552 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296580 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296610 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296640 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296672 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296701 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296818 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296852 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296883 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296917 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296949 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296981 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297012 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297040 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297070 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297101 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297151 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297184 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297293 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297366 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297400 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297459 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297490 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.298737 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.298849 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.298888 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.298952 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.307349 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.307534 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.307662 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.308258 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.308848 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294684 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.294707 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.309852 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295159 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.295764 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.309924 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296041 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296289 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296443 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296816 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.296932 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297343 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.297515 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.298091 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.298587 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.298946 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.310480 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.299074 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.299502 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.299547 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.299474 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.306729 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.306867 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.308178 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.308688 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.310670 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.309796 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.316966 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.316992 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.310944 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.317709 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.317953 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.318092 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.318543 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.318558 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.313088 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.319154 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.319151 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.319368 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.319654 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.320045 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.320182 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.320210 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.320394 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.320515 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.320629 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.320792 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.320929 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.320995 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.321183 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.321448 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.321664 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.321729 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.321939 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.322183 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.322251 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.324605 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.324933 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325105 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325298 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325455 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.325563 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:07:09.810731569 +0000 UTC m=+18.816436750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325590 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325665 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325701 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325806 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325864 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325901 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325926 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325937 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325984 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.325996 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326005 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326028 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326052 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326091 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326125 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326158 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326193 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326221 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326254 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326311 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326346 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326379 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326413 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326446 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326479 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326514 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326546 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326621 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326654 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326686 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326719 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326750 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326781 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326815 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326851 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326886 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326272 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326306 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326511 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326514 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326571 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326708 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326779 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326890 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327119 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.326919 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327335 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327359 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327370 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327432 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327470 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327475 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327508 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327584 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327633 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327677 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327712 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327761 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327767 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327807 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327844 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327878 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327916 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327952 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.327986 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328017 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328042 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328053 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328087 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328120 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328153 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328186 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328223 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328283 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328316 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328353 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328386 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328421 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328459 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328495 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328565 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328614 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328651 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328687 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328725 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328767 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328815 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328851 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329101 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329142 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329177 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329212 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329271 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331248 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331362 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331385 4718 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331408 4718 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331428 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331447 4718 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331466 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331486 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331506 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331527 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331546 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331566 4718 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331585 4718 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331606 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331625 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331645 4718 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331663 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331681 4718 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331698 4718 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331716 4718 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331733 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331749 4718 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331767 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331785 4718 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331826 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331845 4718 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331864 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331881 4718 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331901 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331919 4718 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331938 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331959 4718 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331977 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.331996 4718 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332015 4718 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332035 4718 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332054 4718 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332072 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332091 4718 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332110 4718 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332129 4718 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332148 4718 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332168 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332187 4718 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332211 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332288 4718 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332310 4718 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332328 4718 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332346 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332366 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332386 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332406 4718 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332425 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332448 4718 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332467 4718 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332487 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332505 4718 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332524 4718 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332543 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332561 4718 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332580 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332598 4718 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332616 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332634 4718 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332653 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332671 4718 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332689 4718 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332708 4718 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332727 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332748 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332767 4718 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332786 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332804 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332824 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332844 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332862 4718 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332885 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332905 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.332990 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.333035 4718 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.333053 4718 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.333071 4718 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.333089 4718 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.333117 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.333136 4718 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.333155 4718 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.333173 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.334929 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328247 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328423 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.335641 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328574 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328752 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328933 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.335698 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.328991 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329248 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329255 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329288 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329476 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.329502 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.330359 4718 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52116->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.334212 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.334280 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.334757 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.334790 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.334816 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.335597 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.335856 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52116->192.168.126.11:17697: read: connection reset by peer" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.335909 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.335959 4718 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.336006 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:09.835989082 +0000 UTC m=+18.841694253 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336124 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336523 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336599 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336686 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336741 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336762 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336882 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336917 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336928 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.336992 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.335579 4718 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.337065 4718 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.337074 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.337101 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.337112 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.337128 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:09.837110875 +0000 UTC m=+18.842816046 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.337460 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.337568 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.341937 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.342248 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.342357 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.342432 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.348529 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.349305 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.349816 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.351649 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.351934 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.353080 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.353187 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.353721 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.353806 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.354100 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.354275 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.354500 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.354703 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.354792 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.354991 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.355263 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.355369 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.355559 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.355550 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.355957 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.355987 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.356334 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.356633 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.356678 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.356920 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.356945 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.356927 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.357017 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.357111 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.357251 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.357362 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.357466 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.357482 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.357758 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.357832 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.357897 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.358192 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.358212 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.358608 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.359043 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.359481 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.360364 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.360461 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.360878 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.360925 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.360939 4718 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.361025 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:09.861004811 +0000 UTC m=+18.866710052 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.361152 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.362110 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.362210 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.364146 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.370012 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.370082 4718 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.370184 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:09.870167604 +0000 UTC m=+18.875872765 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.363149 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.363514 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.364772 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.365929 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.369702 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.368411 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.371418 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.372418 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.372476 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.372732 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.372898 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.372888 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.373185 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.373385 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.373391 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.373451 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.374165 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.375076 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.374382 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.374511 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.375188 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.375373 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.375540 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.375668 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.375985 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.376019 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.377615 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.377856 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.378363 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.379406 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.379413 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.379450 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.380145 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.382316 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.383222 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.391150 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.394304 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.397215 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.398628 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.399220 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.399993 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.401115 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.401178 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.401840 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.404983 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.405544 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.406401 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.408148 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.409053 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.409520 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.410642 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.411388 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.411963 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.413344 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.413950 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.415365 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.415548 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.416268 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.417451 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.418330 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.419398 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.420005 4718 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.420155 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.421329 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.422193 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.422752 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.423798 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.425769 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.426181 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.426645 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.427762 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.428461 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.429749 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.430190 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.431299 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.431942 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.431941 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.433001 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.433533 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.433783 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.433952 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.433986 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434043 4718 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434054 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434064 4718 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434075 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434085 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434094 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434103 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434112 4718 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434121 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434130 4718 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434139 4718 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434148 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434156 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434165 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434173 4718 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434182 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434190 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434199 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434208 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434218 4718 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434244 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434255 4718 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434263 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434272 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434281 4718 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434275 4718 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8" exitCode=255 Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434289 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434321 4718 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434445 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434500 4718 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434518 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434532 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434541 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434550 4718 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434559 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434567 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434576 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434586 4718 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434595 4718 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434604 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434619 4718 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434628 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434636 4718 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434645 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434654 4718 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434663 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434672 4718 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434681 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434691 4718 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434700 4718 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434708 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434718 4718 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434727 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434735 4718 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434743 4718 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434751 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434752 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434775 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434760 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434822 4718 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434831 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434842 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434851 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434862 4718 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434870 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434878 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434886 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434895 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434903 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434912 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434919 4718 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434927 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434936 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434955 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434964 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434974 4718 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434986 4718 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.434996 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435004 4718 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435012 4718 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435022 4718 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435030 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435038 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435046 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435054 4718 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435062 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435069 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435078 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435087 4718 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435094 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435103 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435112 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435122 4718 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435130 4718 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435138 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435147 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435155 4718 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435163 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435172 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435181 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435188 4718 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435196 4718 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435204 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435212 4718 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435626 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435640 4718 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435649 4718 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435657 4718 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435666 4718 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435674 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435683 4718 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.435764 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.437404 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.438010 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.439072 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.439674 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.440310 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.451948 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.452699 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.453736 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8"} Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.455036 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.455570 4718 scope.go:117] "RemoveContainer" containerID="f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.463854 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.479185 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.490940 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.500458 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.510343 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.519260 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.584775 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 04:07:09 crc kubenswrapper[4718]: W1206 04:07:09.603676 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-84b7e3ddb3ff2a3381b7484adad9221719bc38b07b5264a3595b382348510ca5 WatchSource:0}: Error finding container 84b7e3ddb3ff2a3381b7484adad9221719bc38b07b5264a3595b382348510ca5: Status 404 returned error can't find the container with id 84b7e3ddb3ff2a3381b7484adad9221719bc38b07b5264a3595b382348510ca5 Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.660405 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 04:07:09 crc kubenswrapper[4718]: W1206 04:07:09.671310 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-a40503d9c339570f2f00aff6bcec0b91d8822250ec271de88498b010ef6c326c WatchSource:0}: Error finding container a40503d9c339570f2f00aff6bcec0b91d8822250ec271de88498b010ef6c326c: Status 404 returned error can't find the container with id a40503d9c339570f2f00aff6bcec0b91d8822250ec271de88498b010ef6c326c Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.678356 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 04:07:09 crc kubenswrapper[4718]: W1206 04:07:09.697800 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-b9f04c60a5722b921b0c037c3f5523651818e8ea986635074ac70288598c82fd WatchSource:0}: Error finding container b9f04c60a5722b921b0c037c3f5523651818e8ea986635074ac70288598c82fd: Status 404 returned error can't find the container with id b9f04c60a5722b921b0c037c3f5523651818e8ea986635074ac70288598c82fd Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.838794 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.839801 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.839864 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.839955 4718 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.840018 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:10.839998747 +0000 UTC m=+19.845703928 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.840099 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:07:10.840086309 +0000 UTC m=+19.845791480 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.840181 4718 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.840220 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:10.840207962 +0000 UTC m=+19.845913143 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.940898 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.940944 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.941076 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.941126 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.941141 4718 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.941185 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:10.941169244 +0000 UTC m=+19.946874415 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.941265 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.941278 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.941288 4718 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:09 crc kubenswrapper[4718]: E1206 04:07:09.941314 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:10.941306607 +0000 UTC m=+19.947011778 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:09 crc kubenswrapper[4718]: I1206 04:07:09.980097 4718 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.437864 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"84b7e3ddb3ff2a3381b7484adad9221719bc38b07b5264a3595b382348510ca5"} Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.442024 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.444079 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e"} Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.444705 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.445635 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7"} Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.445679 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b9f04c60a5722b921b0c037c3f5523651818e8ea986635074ac70288598c82fd"} Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.447650 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a"} Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.447680 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073"} Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.447694 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a40503d9c339570f2f00aff6bcec0b91d8822250ec271de88498b010ef6c326c"} Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.463111 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.486154 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.500168 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.513583 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.530168 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.552140 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.571777 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.584977 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.603061 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.616767 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.641706 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.669192 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.714994 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.729003 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.744425 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-9gdpg"] Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.744761 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9gdpg" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.748610 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.748893 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.749022 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.764207 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.776685 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.792029 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.808158 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.824413 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.840198 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.846353 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.846509 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:07:12.846485336 +0000 UTC m=+21.852190497 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.846552 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.846592 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.846631 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2b4d7e05-3191-4d42-adc8-b07d91e52191-hosts-file\") pod \"node-resolver-9gdpg\" (UID: \"2b4d7e05-3191-4d42-adc8-b07d91e52191\") " pod="openshift-dns/node-resolver-9gdpg" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.846650 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z52pb\" (UniqueName: \"kubernetes.io/projected/2b4d7e05-3191-4d42-adc8-b07d91e52191-kube-api-access-z52pb\") pod \"node-resolver-9gdpg\" (UID: \"2b4d7e05-3191-4d42-adc8-b07d91e52191\") " pod="openshift-dns/node-resolver-9gdpg" Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.846690 4718 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.846770 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:12.846746881 +0000 UTC m=+21.852452042 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.846847 4718 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.846996 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:12.846958376 +0000 UTC m=+21.852663537 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.855080 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.871002 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:10Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.947707 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.947755 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.947792 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2b4d7e05-3191-4d42-adc8-b07d91e52191-hosts-file\") pod \"node-resolver-9gdpg\" (UID: \"2b4d7e05-3191-4d42-adc8-b07d91e52191\") " pod="openshift-dns/node-resolver-9gdpg" Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.947811 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z52pb\" (UniqueName: \"kubernetes.io/projected/2b4d7e05-3191-4d42-adc8-b07d91e52191-kube-api-access-z52pb\") pod \"node-resolver-9gdpg\" (UID: \"2b4d7e05-3191-4d42-adc8-b07d91e52191\") " pod="openshift-dns/node-resolver-9gdpg" Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.947923 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.947955 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.947966 4718 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.948019 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:12.948002729 +0000 UTC m=+21.953707890 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.948051 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2b4d7e05-3191-4d42-adc8-b07d91e52191-hosts-file\") pod \"node-resolver-9gdpg\" (UID: \"2b4d7e05-3191-4d42-adc8-b07d91e52191\") " pod="openshift-dns/node-resolver-9gdpg" Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.948089 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.948138 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.948156 4718 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:10 crc kubenswrapper[4718]: E1206 04:07:10.948257 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:12.948216474 +0000 UTC m=+21.953921635 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:10 crc kubenswrapper[4718]: I1206 04:07:10.965313 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z52pb\" (UniqueName: \"kubernetes.io/projected/2b4d7e05-3191-4d42-adc8-b07d91e52191-kube-api-access-z52pb\") pod \"node-resolver-9gdpg\" (UID: \"2b4d7e05-3191-4d42-adc8-b07d91e52191\") " pod="openshift-dns/node-resolver-9gdpg" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.056439 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9gdpg" Dec 06 04:07:11 crc kubenswrapper[4718]: W1206 04:07:11.073269 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b4d7e05_3191_4d42_adc8_b07d91e52191.slice/crio-69c9634b0e11cde265cd41be25906efdc434a2c632a030a14ababcf189546f44 WatchSource:0}: Error finding container 69c9634b0e11cde265cd41be25906efdc434a2c632a030a14ababcf189546f44: Status 404 returned error can't find the container with id 69c9634b0e11cde265cd41be25906efdc434a2c632a030a14ababcf189546f44 Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.159259 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-pr7fz"] Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.159540 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-jf6pj"] Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.160113 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.160112 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: W1206 04:07:11.162355 4718 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 06 04:07:11 crc kubenswrapper[4718]: E1206 04:07:11.162420 4718 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.163389 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-4r7xc"] Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.163880 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: W1206 04:07:11.165260 4718 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 06 04:07:11 crc kubenswrapper[4718]: E1206 04:07:11.165308 4718 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 04:07:11 crc kubenswrapper[4718]: W1206 04:07:11.165416 4718 reflector.go:561] object-"openshift-multus"/"default-cni-sysctl-allowlist": failed to list *v1.ConfigMap: configmaps "default-cni-sysctl-allowlist" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 06 04:07:11 crc kubenswrapper[4718]: E1206 04:07:11.165481 4718 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"default-cni-sysctl-allowlist\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 04:07:11 crc kubenswrapper[4718]: W1206 04:07:11.165547 4718 reflector.go:561] object-"openshift-machine-config-operator"/"proxy-tls": failed to list *v1.Secret: secrets "proxy-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 06 04:07:11 crc kubenswrapper[4718]: E1206 04:07:11.165560 4718 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"proxy-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"proxy-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.165731 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.165779 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.165972 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.166090 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.166141 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.171172 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.171337 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.174532 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.180707 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.197674 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.218979 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.240678 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.252826 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-var-lib-cni-multus\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.252872 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz2nc\" (UniqueName: \"kubernetes.io/projected/2272fbce-ce87-4f9f-b1c5-f83f262534de-kube-api-access-hz2nc\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.252892 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/87a3c869-d0a2-46cd-ac46-10022d92c7af-rootfs\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.252910 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-system-cni-dir\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.252925 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.252940 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-socket-dir-parent\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.252995 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-system-cni-dir\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253013 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-etc-kubernetes\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253029 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-os-release\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253047 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2272fbce-ce87-4f9f-b1c5-f83f262534de-cni-binary-copy\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253069 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-daemon-config\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253088 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/87a3c869-d0a2-46cd-ac46-10022d92c7af-proxy-tls\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253103 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-cni-dir\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253142 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-run-netns\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253174 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-cnibin\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253203 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/87a3c869-d0a2-46cd-ac46-10022d92c7af-mcd-auth-proxy-config\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253223 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-binary-copy\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253264 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-cnibin\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253299 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-var-lib-kubelet\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253332 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-run-multus-certs\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253361 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-var-lib-cni-bin\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253381 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-hostroot\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253503 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253567 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-conf-dir\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253593 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-os-release\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253644 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-run-k8s-cni-cncf-io\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253662 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bmp2\" (UniqueName: \"kubernetes.io/projected/606ea280-e52e-4894-86cd-454e3f550227-kube-api-access-2bmp2\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.253724 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpbx5\" (UniqueName: \"kubernetes.io/projected/87a3c869-d0a2-46cd-ac46-10022d92c7af-kube-api-access-wpbx5\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.256778 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.276583 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.290705 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.306879 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.319476 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.327416 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.327416 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:11 crc kubenswrapper[4718]: E1206 04:07:11.327571 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.327416 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:11 crc kubenswrapper[4718]: E1206 04:07:11.327715 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:11 crc kubenswrapper[4718]: E1206 04:07:11.327815 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.331453 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.332392 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.337619 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355365 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-cnibin\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355442 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-var-lib-kubelet\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355490 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-run-multus-certs\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355521 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-var-lib-cni-bin\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355545 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-hostroot\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355572 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355577 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-run-multus-certs\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355512 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-cnibin\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355644 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-var-lib-kubelet\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355745 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-conf-dir\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355737 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-hostroot\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355798 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-conf-dir\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355899 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-os-release\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.355992 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-run-k8s-cni-cncf-io\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356018 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bmp2\" (UniqueName: \"kubernetes.io/projected/606ea280-e52e-4894-86cd-454e3f550227-kube-api-access-2bmp2\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356074 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpbx5\" (UniqueName: \"kubernetes.io/projected/87a3c869-d0a2-46cd-ac46-10022d92c7af-kube-api-access-wpbx5\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356095 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-var-lib-cni-multus\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356113 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz2nc\" (UniqueName: \"kubernetes.io/projected/2272fbce-ce87-4f9f-b1c5-f83f262534de-kube-api-access-hz2nc\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356132 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/87a3c869-d0a2-46cd-ac46-10022d92c7af-rootfs\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356154 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-system-cni-dir\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356173 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356210 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-socket-dir-parent\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356247 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-system-cni-dir\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356279 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-os-release\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356300 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2272fbce-ce87-4f9f-b1c5-f83f262534de-cni-binary-copy\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356332 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-etc-kubernetes\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356354 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-system-cni-dir\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356395 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-os-release\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356363 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-daemon-config\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356353 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356465 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/87a3c869-d0a2-46cd-ac46-10022d92c7af-proxy-tls\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356502 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-cni-dir\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356525 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-run-netns\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356559 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-cnibin\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356585 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/87a3c869-d0a2-46cd-ac46-10022d92c7af-mcd-auth-proxy-config\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356619 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-binary-copy\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356621 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-os-release\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356270 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-var-lib-cni-bin\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356708 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-var-lib-cni-multus\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356726 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-etc-kubernetes\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356733 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-socket-dir-parent\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356872 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-system-cni-dir\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356902 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-cnibin\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356940 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/87a3c869-d0a2-46cd-ac46-10022d92c7af-rootfs\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356977 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-cni-dir\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.356984 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/606ea280-e52e-4894-86cd-454e3f550227-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.357007 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-run-k8s-cni-cncf-io\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.357042 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2272fbce-ce87-4f9f-b1c5-f83f262534de-host-run-netns\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.357786 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2272fbce-ce87-4f9f-b1c5-f83f262534de-multus-daemon-config\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.358188 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/87a3c869-d0a2-46cd-ac46-10022d92c7af-mcd-auth-proxy-config\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.405479 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.405616 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bmp2\" (UniqueName: \"kubernetes.io/projected/606ea280-e52e-4894-86cd-454e3f550227-kube-api-access-2bmp2\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.415405 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz2nc\" (UniqueName: \"kubernetes.io/projected/2272fbce-ce87-4f9f-b1c5-f83f262534de-kube-api-access-hz2nc\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.422718 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpbx5\" (UniqueName: \"kubernetes.io/projected/87a3c869-d0a2-46cd-ac46-10022d92c7af-kube-api-access-wpbx5\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.426589 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.445803 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.451366 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9gdpg" event={"ID":"2b4d7e05-3191-4d42-adc8-b07d91e52191","Type":"ContainerStarted","Data":"d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa"} Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.451403 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9gdpg" event={"ID":"2b4d7e05-3191-4d42-adc8-b07d91e52191","Type":"ContainerStarted","Data":"69c9634b0e11cde265cd41be25906efdc434a2c632a030a14ababcf189546f44"} Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.462320 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.476044 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.493092 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.511523 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.524603 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.537969 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.551523 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.567609 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.577205 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6nfdl"] Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.578079 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.582382 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.582471 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.582736 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.582810 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.582941 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.583775 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.585500 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.585573 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.600962 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.613358 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.634127 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.653994 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658396 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-systemd-units\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658435 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-systemd\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658455 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-env-overrides\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658488 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-ovn\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658514 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-bin\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658572 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnpgw\" (UniqueName: \"kubernetes.io/projected/e9ed5db1-456b-4891-a750-809728e9c0a5-kube-api-access-qnpgw\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658629 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-kubelet\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658798 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-netns\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658956 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-openvswitch\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.658988 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-log-socket\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659026 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-config\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659147 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659199 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-etc-openvswitch\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659264 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-netd\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659302 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-slash\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659333 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-var-lib-openvswitch\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659386 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-node-log\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659417 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-script-lib\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659450 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e9ed5db1-456b-4891-a750-809728e9c0a5-ovn-node-metrics-cert\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.659518 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-ovn-kubernetes\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.699184 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.731547 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.756418 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760129 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-slash\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760195 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-var-lib-openvswitch\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760223 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-netd\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760290 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-node-log\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760313 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-script-lib\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760336 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e9ed5db1-456b-4891-a750-809728e9c0a5-ovn-node-metrics-cert\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760334 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-slash\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760374 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-var-lib-openvswitch\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760371 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-ovn-kubernetes\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760431 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-ovn-kubernetes\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760484 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-netd\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760524 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-systemd-units\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760498 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-systemd-units\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760655 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-systemd\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760693 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-env-overrides\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760699 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-systemd\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760653 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-node-log\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760732 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-ovn\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760805 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-bin\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760848 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-netns\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760874 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-openvswitch\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760880 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-bin\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760842 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-ovn\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760900 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-log-socket\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760928 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-netns\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760927 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnpgw\" (UniqueName: \"kubernetes.io/projected/e9ed5db1-456b-4891-a750-809728e9c0a5-kube-api-access-qnpgw\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760965 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-openvswitch\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760980 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-kubelet\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.760996 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-log-socket\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.761016 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-config\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.761067 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-etc-openvswitch\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.761093 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.761162 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-script-lib\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.761167 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.761226 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-env-overrides\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.761285 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-etc-openvswitch\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.761303 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-kubelet\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.761810 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-config\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.771288 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e9ed5db1-456b-4891-a750-809728e9c0a5-ovn-node-metrics-cert\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.773132 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.778399 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnpgw\" (UniqueName: \"kubernetes.io/projected/e9ed5db1-456b-4891-a750-809728e9c0a5-kube-api-access-qnpgw\") pod \"ovnkube-node-6nfdl\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.788958 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.812594 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.827541 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.841951 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.858879 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.873287 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.891740 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.891722 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.912080 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.925770 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.941777 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.959432 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.972907 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:11 crc kubenswrapper[4718]: I1206 04:07:11.989018 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.018225 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.034965 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.049481 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.064767 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.078112 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.100070 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.119382 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.134918 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.153457 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.196765 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.214435 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.247481 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.248328 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.260402 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/87a3c869-d0a2-46cd-ac46-10022d92c7af-proxy-tls\") pod \"machine-config-daemon-pr7fz\" (UID: \"87a3c869-d0a2-46cd-ac46-10022d92c7af\") " pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.293615 4718 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.295103 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.295159 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.295171 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.295332 4718 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.302023 4718 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.302533 4718 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.303785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.303827 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.303838 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.303856 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.303870 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.316646 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.323980 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.324011 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.324021 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.324038 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.324050 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.336532 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.340656 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.340683 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.340692 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.340706 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.340714 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.356218 4718 configmap.go:193] Couldn't get configMap openshift-multus/default-cni-sysctl-allowlist: failed to sync configmap cache: timed out waiting for the condition Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.356321 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-sysctl-allowlist podName:606ea280-e52e-4894-86cd-454e3f550227 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:12.856302656 +0000 UTC m=+21.862007817 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-sysctl-allowlist" (UniqueName: "kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-sysctl-allowlist") pod "multus-additional-cni-plugins-jf6pj" (UID: "606ea280-e52e-4894-86cd-454e3f550227") : failed to sync configmap cache: timed out waiting for the condition Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.356378 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.357339 4718 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.357428 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2272fbce-ce87-4f9f-b1c5-f83f262534de-cni-binary-copy podName:2272fbce-ce87-4f9f-b1c5-f83f262534de nodeName:}" failed. No retries permitted until 2025-12-06 04:07:12.857411237 +0000 UTC m=+21.863116398 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/2272fbce-ce87-4f9f-b1c5-f83f262534de-cni-binary-copy") pod "multus-4r7xc" (UID: "2272fbce-ce87-4f9f-b1c5-f83f262534de") : failed to sync configmap cache: timed out waiting for the condition Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.357349 4718 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.357463 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-binary-copy podName:606ea280-e52e-4894-86cd-454e3f550227 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:12.857457698 +0000 UTC m=+21.863162859 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-binary-copy") pod "multus-additional-cni-plugins-jf6pj" (UID: "606ea280-e52e-4894-86cd-454e3f550227") : failed to sync configmap cache: timed out waiting for the condition Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.369501 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.369551 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.369563 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.369581 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.369593 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.382424 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.383399 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.385747 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.385802 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.385819 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.385842 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.385860 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.398415 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.398546 4718 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.400818 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.400856 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.400869 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.400894 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.400910 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.433814 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.455294 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe" exitCode=0 Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.455369 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.455397 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"1a7fb23f2bc1e4d7ea3c66d92ddb9fc20b39c4f38194f736af44dfb9315241bc"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.457141 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.458322 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"9d77b5a5534471fd7ed798320b9c4c2668eb33801d3579d7c04cce296d33bf65"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.485911 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.500156 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.503520 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.503566 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.503582 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.503603 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.503618 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.529403 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.545221 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.555856 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.559906 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.581031 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.582943 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.598163 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.606561 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.606627 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.606645 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.606663 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.606677 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.614114 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.626681 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.641027 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.684858 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.708910 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.708942 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.708958 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.708970 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.708979 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.717721 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.757987 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.803557 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.812148 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.812190 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.812201 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.812220 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.812253 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.838609 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.873263 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.874457 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.874527 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-binary-copy\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.874551 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.874582 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:07:16.874567015 +0000 UTC m=+25.880272176 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.874616 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.874635 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.874665 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2272fbce-ce87-4f9f-b1c5-f83f262534de-cni-binary-copy\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.874847 4718 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.874925 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:16.874909341 +0000 UTC m=+25.880614492 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.874857 4718 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.875029 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:16.875006883 +0000 UTC m=+25.880712104 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.875319 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.875475 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2272fbce-ce87-4f9f-b1c5-f83f262534de-cni-binary-copy\") pod \"multus-4r7xc\" (UID: \"2272fbce-ce87-4f9f-b1c5-f83f262534de\") " pod="openshift-multus/multus-4r7xc" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.875562 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/606ea280-e52e-4894-86cd-454e3f550227-cni-binary-copy\") pod \"multus-additional-cni-plugins-jf6pj\" (UID: \"606ea280-e52e-4894-86cd-454e3f550227\") " pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.914849 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.915275 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.915288 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.915308 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.915320 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:12Z","lastTransitionTime":"2025-12-06T04:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.920335 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.958037 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.974525 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.975578 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.975620 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.975769 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.975780 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.975812 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.975827 4718 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.975881 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:16.975862423 +0000 UTC m=+25.981567594 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.975786 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.975916 4718 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:12 crc kubenswrapper[4718]: E1206 04:07:12.975969 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:16.975954705 +0000 UTC m=+25.981659866 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:12 crc kubenswrapper[4718]: W1206 04:07:12.989394 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod606ea280_e52e_4894_86cd_454e3f550227.slice/crio-daea1382451df03d8a40552820dd72444ff1848e21d2f298ffc25b441f1df15e WatchSource:0}: Error finding container daea1382451df03d8a40552820dd72444ff1848e21d2f298ffc25b441f1df15e: Status 404 returned error can't find the container with id daea1382451df03d8a40552820dd72444ff1848e21d2f298ffc25b441f1df15e Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.991062 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4r7xc" Dec 06 04:07:12 crc kubenswrapper[4718]: I1206 04:07:12.997973 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:12Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: W1206 04:07:13.010677 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2272fbce_ce87_4f9f_b1c5_f83f262534de.slice/crio-6a2dc8fc6095c60040cff9457ac0728df64d30cc31fc3cb202ca56453af5c61e WatchSource:0}: Error finding container 6a2dc8fc6095c60040cff9457ac0728df64d30cc31fc3cb202ca56453af5c61e: Status 404 returned error can't find the container with id 6a2dc8fc6095c60040cff9457ac0728df64d30cc31fc3cb202ca56453af5c61e Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.018156 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.018351 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.018673 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.018769 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.018908 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.038248 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.075552 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.104542 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.111449 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.122340 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.123763 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.123832 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.123849 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.123873 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.123889 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.132185 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.179058 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.218468 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.227754 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.227818 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.227836 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.227863 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.227878 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.257193 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.295979 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.327778 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.327807 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.327893 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:13 crc kubenswrapper[4718]: E1206 04:07:13.327951 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:13 crc kubenswrapper[4718]: E1206 04:07:13.328077 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:13 crc kubenswrapper[4718]: E1206 04:07:13.328206 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.330346 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.330399 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.330418 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.330442 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.330459 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.338304 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.392708 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.426042 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.433527 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.433589 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.433611 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.433637 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.433655 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.464621 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.464663 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.464675 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.464688 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.466285 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.466345 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.467439 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4r7xc" event={"ID":"2272fbce-ce87-4f9f-b1c5-f83f262534de","Type":"ContainerStarted","Data":"6a2dc8fc6095c60040cff9457ac0728df64d30cc31fc3cb202ca56453af5c61e"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.468647 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" event={"ID":"606ea280-e52e-4894-86cd-454e3f550227","Type":"ContainerStarted","Data":"daea1382451df03d8a40552820dd72444ff1848e21d2f298ffc25b441f1df15e"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.473995 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.504705 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.520058 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-fxdkw"] Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.520375 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.536264 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.536428 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.536567 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.536698 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.536775 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.537611 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.546374 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.566347 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.581208 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10ed0065-fa7b-4190-b140-ddcf6c368ef9-host\") pod \"node-ca-fxdkw\" (UID: \"10ed0065-fa7b-4190-b140-ddcf6c368ef9\") " pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.581281 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljm9f\" (UniqueName: \"kubernetes.io/projected/10ed0065-fa7b-4190-b140-ddcf6c368ef9-kube-api-access-ljm9f\") pod \"node-ca-fxdkw\" (UID: \"10ed0065-fa7b-4190-b140-ddcf6c368ef9\") " pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.581311 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/10ed0065-fa7b-4190-b140-ddcf6c368ef9-serviceca\") pod \"node-ca-fxdkw\" (UID: \"10ed0065-fa7b-4190-b140-ddcf6c368ef9\") " pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.591696 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.607485 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.639300 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.639343 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.639357 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.639374 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.639387 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.654345 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.682616 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljm9f\" (UniqueName: \"kubernetes.io/projected/10ed0065-fa7b-4190-b140-ddcf6c368ef9-kube-api-access-ljm9f\") pod \"node-ca-fxdkw\" (UID: \"10ed0065-fa7b-4190-b140-ddcf6c368ef9\") " pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.682664 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/10ed0065-fa7b-4190-b140-ddcf6c368ef9-serviceca\") pod \"node-ca-fxdkw\" (UID: \"10ed0065-fa7b-4190-b140-ddcf6c368ef9\") " pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.682717 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10ed0065-fa7b-4190-b140-ddcf6c368ef9-host\") pod \"node-ca-fxdkw\" (UID: \"10ed0065-fa7b-4190-b140-ddcf6c368ef9\") " pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.682790 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10ed0065-fa7b-4190-b140-ddcf6c368ef9-host\") pod \"node-ca-fxdkw\" (UID: \"10ed0065-fa7b-4190-b140-ddcf6c368ef9\") " pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.683711 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/10ed0065-fa7b-4190-b140-ddcf6c368ef9-serviceca\") pod \"node-ca-fxdkw\" (UID: \"10ed0065-fa7b-4190-b140-ddcf6c368ef9\") " pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.698477 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.726504 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljm9f\" (UniqueName: \"kubernetes.io/projected/10ed0065-fa7b-4190-b140-ddcf6c368ef9-kube-api-access-ljm9f\") pod \"node-ca-fxdkw\" (UID: \"10ed0065-fa7b-4190-b140-ddcf6c368ef9\") " pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.743170 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.743766 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.743785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.743813 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.743831 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.753450 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.796653 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.833873 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.840048 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fxdkw" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.847341 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.847385 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.847395 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.847411 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.847421 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.873905 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.917161 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.949860 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.949908 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.949921 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.949939 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.949951 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:13Z","lastTransitionTime":"2025-12-06T04:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.954268 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:13 crc kubenswrapper[4718]: I1206 04:07:13.996621 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.037818 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.052285 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.052328 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.052342 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.052361 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.052375 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.077591 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.117201 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.154478 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.154521 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.154533 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.154550 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.154560 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.158458 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.194747 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.251605 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.256792 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.256824 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.256834 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.256847 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.256855 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.275755 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.314541 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.356433 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.358871 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.358901 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.358914 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.358930 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.358941 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.397402 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.461348 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.461400 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.461411 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.461427 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.461440 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.472538 4718 generic.go:334] "Generic (PLEG): container finished" podID="606ea280-e52e-4894-86cd-454e3f550227" containerID="e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7" exitCode=0 Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.472634 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" event={"ID":"606ea280-e52e-4894-86cd-454e3f550227","Type":"ContainerDied","Data":"e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.474449 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4r7xc" event={"ID":"2272fbce-ce87-4f9f-b1c5-f83f262534de","Type":"ContainerStarted","Data":"7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.476221 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fxdkw" event={"ID":"10ed0065-fa7b-4190-b140-ddcf6c368ef9","Type":"ContainerStarted","Data":"656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.476266 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fxdkw" event={"ID":"10ed0065-fa7b-4190-b140-ddcf6c368ef9","Type":"ContainerStarted","Data":"27ff9d6d899f74041ce8621846d9940882f317f7a8e2961992ee4d3c7df2e155"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.480311 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.480377 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.495522 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.509586 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.525435 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.555523 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.563779 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.563804 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.563817 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.563832 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.563842 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.606587 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.633335 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.665568 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.665610 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.665621 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.665637 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.665648 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.673014 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.714721 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.753120 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.767890 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.767915 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.767923 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.767936 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.767946 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.796362 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.839735 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.870335 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.870372 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.870383 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.870399 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.870410 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.877858 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.918195 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.956831 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.972581 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.972615 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.972625 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.972640 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.972651 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:14Z","lastTransitionTime":"2025-12-06T04:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:14 crc kubenswrapper[4718]: I1206 04:07:14.997897 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:14Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.037813 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.076027 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.076069 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.076082 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.076098 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.076112 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.077787 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.114936 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.155450 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.178212 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.178258 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.178267 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.178282 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.178291 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.200513 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.237337 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.278847 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.281425 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.281466 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.281474 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.281498 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.281506 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.314742 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.327700 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:15 crc kubenswrapper[4718]: E1206 04:07:15.327819 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.328035 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:15 crc kubenswrapper[4718]: E1206 04:07:15.328089 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.328127 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:15 crc kubenswrapper[4718]: E1206 04:07:15.328163 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.363025 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.383818 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.383855 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.383866 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.383881 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.383892 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.394403 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.434867 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.477589 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.485416 4718 generic.go:334] "Generic (PLEG): container finished" podID="606ea280-e52e-4894-86cd-454e3f550227" containerID="4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343" exitCode=0 Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.485543 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" event={"ID":"606ea280-e52e-4894-86cd-454e3f550227","Type":"ContainerDied","Data":"4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.485681 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.485742 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.485760 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.485815 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.485833 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.527348 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.570782 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.587960 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.588005 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.588018 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.588037 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.588050 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.605815 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.635219 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.674139 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.690098 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.690150 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.690161 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.690178 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.690190 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.716708 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.755459 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.792591 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.792635 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.792646 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.792665 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.792677 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.797458 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.835444 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.874362 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.894490 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.894543 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.894557 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.894576 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.894587 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.915575 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.956348 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.997292 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.997332 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.997341 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.997357 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:15 crc kubenswrapper[4718]: I1206 04:07:15.997370 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:15Z","lastTransitionTime":"2025-12-06T04:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.002143 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:15Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.036663 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.085767 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.100051 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.100105 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.100117 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.100134 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.100147 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:16Z","lastTransitionTime":"2025-12-06T04:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.203016 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.203081 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.203098 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.203123 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.203140 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:16Z","lastTransitionTime":"2025-12-06T04:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.305827 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.305869 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.305880 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.305896 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.305908 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:16Z","lastTransitionTime":"2025-12-06T04:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.408737 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.408788 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.408799 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.408817 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.408828 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:16Z","lastTransitionTime":"2025-12-06T04:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.490785 4718 generic.go:334] "Generic (PLEG): container finished" podID="606ea280-e52e-4894-86cd-454e3f550227" containerID="298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e" exitCode=0 Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.490888 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" event={"ID":"606ea280-e52e-4894-86cd-454e3f550227","Type":"ContainerDied","Data":"298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.496352 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.512368 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.512549 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.512752 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.512981 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.513078 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:16Z","lastTransitionTime":"2025-12-06T04:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.518530 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.531704 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.546776 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.559099 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.578536 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.589996 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.602371 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.613874 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.615845 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.615885 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.615898 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.615915 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.615925 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:16Z","lastTransitionTime":"2025-12-06T04:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.628589 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.645580 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.667482 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.687997 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.705882 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.718420 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.718478 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.718494 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.718516 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.718533 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:16Z","lastTransitionTime":"2025-12-06T04:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.719961 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:16Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.823958 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.823997 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.824005 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.824022 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.824035 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:16Z","lastTransitionTime":"2025-12-06T04:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.916629 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.916823 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.916863 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:16 crc kubenswrapper[4718]: E1206 04:07:16.916897 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:07:24.916864995 +0000 UTC m=+33.922570166 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:07:16 crc kubenswrapper[4718]: E1206 04:07:16.917023 4718 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:16 crc kubenswrapper[4718]: E1206 04:07:16.917174 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:24.91714342 +0000 UTC m=+33.922848621 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:16 crc kubenswrapper[4718]: E1206 04:07:16.917033 4718 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:16 crc kubenswrapper[4718]: E1206 04:07:16.917320 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:24.917297603 +0000 UTC m=+33.923002804 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.926248 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.926282 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.926291 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.926304 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:16 crc kubenswrapper[4718]: I1206 04:07:16.926332 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:16Z","lastTransitionTime":"2025-12-06T04:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.017663 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.017728 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.017907 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.017931 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.017930 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.017987 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.018005 4718 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.017947 4718 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.018083 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:25.018060082 +0000 UTC m=+34.023765313 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.018180 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:25.018159014 +0000 UTC m=+34.023864185 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.029719 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.029770 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.029785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.029807 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.029822 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.132469 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.132559 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.132588 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.132621 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.132645 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.235903 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.235971 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.235993 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.236018 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.236037 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.328220 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.328300 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.328220 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.328448 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.328543 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:17 crc kubenswrapper[4718]: E1206 04:07:17.328703 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.338826 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.338917 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.338941 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.338969 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.338992 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.442184 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.442253 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.442272 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.442295 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.442309 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.501292 4718 generic.go:334] "Generic (PLEG): container finished" podID="606ea280-e52e-4894-86cd-454e3f550227" containerID="88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95" exitCode=0 Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.501362 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" event={"ID":"606ea280-e52e-4894-86cd-454e3f550227","Type":"ContainerDied","Data":"88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.524397 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.534480 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.544727 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.544780 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.544797 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.544827 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.544845 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.546008 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.560292 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.594863 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.607953 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.620920 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.634179 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.646080 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.647944 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.647987 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.647998 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.648016 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.648027 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.658033 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.672298 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.683904 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.700453 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.714739 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:17Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.750607 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.750646 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.750656 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.750672 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.750681 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.853644 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.853696 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.853707 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.853727 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.853739 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.957787 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.957999 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.958112 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.958271 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:17 crc kubenswrapper[4718]: I1206 04:07:17.958423 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:17Z","lastTransitionTime":"2025-12-06T04:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.061875 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.061927 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.061940 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.061961 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.061974 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.165135 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.165206 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.165288 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.165323 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.165348 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.270486 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.270533 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.270548 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.270570 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.270585 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.373557 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.373618 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.373635 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.373657 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.373673 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.476105 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.476169 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.476187 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.476212 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.476263 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.509225 4718 generic.go:334] "Generic (PLEG): container finished" podID="606ea280-e52e-4894-86cd-454e3f550227" containerID="0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13" exitCode=0 Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.509341 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" event={"ID":"606ea280-e52e-4894-86cd-454e3f550227","Type":"ContainerDied","Data":"0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.520992 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.522206 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.522365 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.522418 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.532708 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.556527 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.559305 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.560058 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.577947 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.580081 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.580129 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.580145 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.580165 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.580179 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.592174 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.605160 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.621079 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.634889 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.647375 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.662728 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.676520 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.682461 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.682493 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.682506 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.682525 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.682536 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.692734 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.707777 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.723575 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.733450 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.743821 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.759731 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.775474 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.785747 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.785880 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.785962 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.786129 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.786333 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.789909 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.811614 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.831620 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.844966 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.866671 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.883718 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.890910 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.890951 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.890965 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.890983 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.890995 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.900796 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.917721 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.932838 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.946375 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.963436 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:18Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.993056 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.993095 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.993105 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.993120 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:18 crc kubenswrapper[4718]: I1206 04:07:18.993132 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:18Z","lastTransitionTime":"2025-12-06T04:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.096252 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.096281 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.096290 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.096303 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.096312 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:19Z","lastTransitionTime":"2025-12-06T04:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.198725 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.198797 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.198822 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.198849 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.198871 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:19Z","lastTransitionTime":"2025-12-06T04:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.302046 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.302092 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.302104 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.302122 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.302135 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:19Z","lastTransitionTime":"2025-12-06T04:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.327391 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.327393 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:19 crc kubenswrapper[4718]: E1206 04:07:19.327521 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.327352 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:19 crc kubenswrapper[4718]: E1206 04:07:19.327594 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:19 crc kubenswrapper[4718]: E1206 04:07:19.327771 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.404272 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.404319 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.404329 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.404342 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.404351 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:19Z","lastTransitionTime":"2025-12-06T04:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.506571 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.506647 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.506677 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.506707 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.506726 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:19Z","lastTransitionTime":"2025-12-06T04:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.529730 4718 generic.go:334] "Generic (PLEG): container finished" podID="606ea280-e52e-4894-86cd-454e3f550227" containerID="f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065" exitCode=0 Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.529793 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" event={"ID":"606ea280-e52e-4894-86cd-454e3f550227","Type":"ContainerDied","Data":"f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.542852 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.606884 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.609324 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.609353 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.609364 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.609379 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.609389 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:19Z","lastTransitionTime":"2025-12-06T04:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.625437 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.640964 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.666459 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.687452 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.704677 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.713836 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.713868 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.713883 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.713900 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.713911 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:19Z","lastTransitionTime":"2025-12-06T04:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.733912 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.750934 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.766373 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.782030 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.795959 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.810056 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.816450 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.816486 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.816495 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.816512 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.816523 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:19Z","lastTransitionTime":"2025-12-06T04:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.824771 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:19Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.919384 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.919447 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.919458 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.919480 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:19 crc kubenswrapper[4718]: I1206 04:07:19.919495 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:19Z","lastTransitionTime":"2025-12-06T04:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.021638 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.021677 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.021688 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.021702 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.021744 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.125598 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.125660 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.125671 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.125692 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.125704 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.227814 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.227846 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.227856 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.227871 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.227883 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.342227 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.342279 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.342288 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.342300 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.342310 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.444507 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.444537 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.444545 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.444558 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.444567 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.536335 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" event={"ID":"606ea280-e52e-4894-86cd-454e3f550227","Type":"ContainerStarted","Data":"77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.546454 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.546499 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.546512 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.546530 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.546542 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.550632 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.559912 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.571495 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.583079 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.591063 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.607048 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.620187 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.632558 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.649363 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.649407 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.649419 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.649436 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.649448 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.651032 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.664969 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.677428 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.691491 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.707411 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.724323 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.753179 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.753214 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.753226 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.753271 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.753284 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.850864 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.855924 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.856003 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.856035 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.856113 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.856140 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.866345 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.896969 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.913860 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.928046 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.947983 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.959295 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.959349 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.959362 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.959380 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.959394 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:20Z","lastTransitionTime":"2025-12-06T04:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.959569 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.971760 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.984489 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:20 crc kubenswrapper[4718]: I1206 04:07:20.994743 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:20Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.008254 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.021446 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.036300 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.050032 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.062379 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.062439 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.062454 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.062480 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.062494 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.068938 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.165741 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.165811 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.165840 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.165874 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.165895 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.268886 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.268988 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.269008 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.269034 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.269055 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.327525 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.327557 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:21 crc kubenswrapper[4718]: E1206 04:07:21.327744 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.328225 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:21 crc kubenswrapper[4718]: E1206 04:07:21.328497 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:21 crc kubenswrapper[4718]: E1206 04:07:21.328984 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.348816 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.364792 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.371620 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.371648 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.371657 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.371673 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.371684 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.388124 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.426639 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.443402 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.456719 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.474351 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.474621 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.474880 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.474990 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.475089 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.474677 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.499946 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.516167 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.535010 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.541940 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/0.log" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.544287 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b" exitCode=1 Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.545194 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.546206 4718 scope.go:117] "RemoveContainer" containerID="ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.557386 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.579920 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.580294 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.580338 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.580356 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.580382 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.580401 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.597922 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.616001 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.633288 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.651503 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.670385 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.682147 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.682210 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.682260 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.682294 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.682312 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.688909 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.706491 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.726174 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.744308 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.758537 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.770976 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.783563 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.784961 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.785000 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.785012 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.785027 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.785037 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.794723 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.808334 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.819573 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.839677 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"message\\\":\\\"nNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:20.891153 5982 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:20.891882 5982 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:20.891920 5982 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:20.891951 5982 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:20.892000 5982 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:20.892034 5982 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:20.892047 5982 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:20.892050 5982 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:20.892079 5982 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:20.892082 5982 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 04:07:20.892085 5982 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 04:07:20.892111 5982 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:20.892131 5982 factory.go:656] Stopping watch factory\\\\nI1206 04:07:20.892144 5982 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:21Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.887249 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.887278 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.887286 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.887299 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.887308 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.989923 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.989963 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.989978 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.989994 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:21 crc kubenswrapper[4718]: I1206 04:07:21.990007 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:21Z","lastTransitionTime":"2025-12-06T04:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.092116 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.092183 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.092194 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.092211 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.092225 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.194784 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.194837 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.194850 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.194873 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.194886 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.297869 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.297917 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.297931 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.297949 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.297962 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.400801 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.400862 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.400877 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.400923 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.400936 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.503573 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.503618 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.503629 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.503645 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.503657 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.549841 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/0.log" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.552642 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.553168 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.567608 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.577812 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.587705 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.605964 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.606007 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.606016 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.606030 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.606039 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.616679 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"message\\\":\\\"nNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:20.891153 5982 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:20.891882 5982 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:20.891920 5982 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:20.891951 5982 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:20.892000 5982 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:20.892034 5982 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:20.892047 5982 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:20.892050 5982 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:20.892079 5982 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:20.892082 5982 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 04:07:20.892085 5982 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 04:07:20.892111 5982 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:20.892131 5982 factory.go:656] Stopping watch factory\\\\nI1206 04:07:20.892144 5982 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.632327 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.643765 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.657014 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.657065 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.657082 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.657105 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.657123 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.657580 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: E1206 04:07:22.671100 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.672935 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.674878 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.674917 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.674932 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.674949 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.674961 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: E1206 04:07:22.689284 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.695531 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.697643 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.697676 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.697686 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.697705 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.697716 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.710817 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: E1206 04:07:22.713547 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.718177 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.718277 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.718319 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.718350 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.718373 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.727672 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: E1206 04:07:22.732396 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.735553 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.735585 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.735597 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.735616 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.735629 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.744773 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: E1206 04:07:22.748952 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: E1206 04:07:22.749320 4718 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.751126 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.751162 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.751172 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.751186 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.751197 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.759143 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.775030 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:22Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.854035 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.854078 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.854092 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.854113 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.854129 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.956560 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.956599 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.956607 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.956621 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:22 crc kubenswrapper[4718]: I1206 04:07:22.956631 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:22Z","lastTransitionTime":"2025-12-06T04:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.059927 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.059985 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.060003 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.060027 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.060048 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.162383 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.162442 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.162456 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.162476 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.162492 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.191875 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs"] Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.192592 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.195514 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.199646 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.220808 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.238313 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.257162 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.264525 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.264553 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.264565 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.264580 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.264591 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.280172 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.281850 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a81cbde5-9395-418f-a205-279a33268d9e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.281920 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a81cbde5-9395-418f-a205-279a33268d9e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.281937 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a81cbde5-9395-418f-a205-279a33268d9e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.281982 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xch99\" (UniqueName: \"kubernetes.io/projected/a81cbde5-9395-418f-a205-279a33268d9e-kube-api-access-xch99\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.294942 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.313490 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.327480 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.327525 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.327554 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:23 crc kubenswrapper[4718]: E1206 04:07:23.327669 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:23 crc kubenswrapper[4718]: E1206 04:07:23.327864 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:23 crc kubenswrapper[4718]: E1206 04:07:23.327983 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.341741 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.355475 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.367538 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.367583 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.367598 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.367619 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.367722 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.369868 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.380053 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.382965 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xch99\" (UniqueName: \"kubernetes.io/projected/a81cbde5-9395-418f-a205-279a33268d9e-kube-api-access-xch99\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.383021 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a81cbde5-9395-418f-a205-279a33268d9e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.383125 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a81cbde5-9395-418f-a205-279a33268d9e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.383157 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a81cbde5-9395-418f-a205-279a33268d9e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.384078 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a81cbde5-9395-418f-a205-279a33268d9e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.384442 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a81cbde5-9395-418f-a205-279a33268d9e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.392953 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a81cbde5-9395-418f-a205-279a33268d9e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.398710 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.411705 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xch99\" (UniqueName: \"kubernetes.io/projected/a81cbde5-9395-418f-a205-279a33268d9e-kube-api-access-xch99\") pod \"ovnkube-control-plane-749d76644c-7d4cs\" (UID: \"a81cbde5-9395-418f-a205-279a33268d9e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.411766 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.425027 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.436774 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.466659 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"message\\\":\\\"nNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:20.891153 5982 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:20.891882 5982 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:20.891920 5982 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:20.891951 5982 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:20.892000 5982 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:20.892034 5982 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:20.892047 5982 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:20.892050 5982 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:20.892079 5982 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:20.892082 5982 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 04:07:20.892085 5982 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 04:07:20.892111 5982 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:20.892131 5982 factory.go:656] Stopping watch factory\\\\nI1206 04:07:20.892144 5982 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.470259 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.470288 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.470298 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.470313 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.470349 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.512702 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.565929 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" event={"ID":"a81cbde5-9395-418f-a205-279a33268d9e","Type":"ContainerStarted","Data":"bdb5044c2855df4f404d5e26a1cb4a2900f1e841f746ea42fea5fd5ece0fe9a1"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.568411 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/1.log" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.569590 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/0.log" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.571931 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.571958 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.571967 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.571979 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.571987 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.572751 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426" exitCode=1 Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.572782 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.572814 4718 scope.go:117] "RemoveContainer" containerID="ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.573368 4718 scope.go:117] "RemoveContainer" containerID="792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426" Dec 06 04:07:23 crc kubenswrapper[4718]: E1206 04:07:23.573521 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.590931 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.606877 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.618122 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.629009 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.640305 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.649086 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.664391 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab421d33055dfbb24059608ca931c11398158e2b8a345876a25b218ca5e2404b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"message\\\":\\\"nNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:20.891153 5982 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:20.891882 5982 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:20.891920 5982 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:20.891951 5982 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:20.892000 5982 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:20.892034 5982 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:20.892047 5982 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:20.892050 5982 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:20.892079 5982 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:20.892082 5982 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 04:07:20.892085 5982 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 04:07:20.892111 5982 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:20.892131 5982 factory.go:656] Stopping watch factory\\\\nI1206 04:07:20.892144 5982 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.673690 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.673718 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.673728 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.673742 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.673753 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.682494 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.698249 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.712090 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.725367 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.738658 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.750190 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.760611 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.772077 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:23Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.775594 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.775636 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.775646 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.775664 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.775674 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.877906 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.877950 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.877967 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.877982 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.877993 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.980885 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.981209 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.981228 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.981289 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:23 crc kubenswrapper[4718]: I1206 04:07:23.981306 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:23Z","lastTransitionTime":"2025-12-06T04:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.084589 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.084653 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.084669 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.084692 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.084709 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:24Z","lastTransitionTime":"2025-12-06T04:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.187405 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.187452 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.187465 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.187482 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.187494 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:24Z","lastTransitionTime":"2025-12-06T04:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.290290 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.290350 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.290373 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.290398 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.290416 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:24Z","lastTransitionTime":"2025-12-06T04:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.392507 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.392538 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.392547 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.392562 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.392573 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:24Z","lastTransitionTime":"2025-12-06T04:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.495044 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.495091 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.495107 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.495128 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.495188 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:24Z","lastTransitionTime":"2025-12-06T04:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.576643 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/1.log" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.580660 4718 scope.go:117] "RemoveContainer" containerID="792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426" Dec 06 04:07:24 crc kubenswrapper[4718]: E1206 04:07:24.580808 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.582250 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" event={"ID":"a81cbde5-9395-418f-a205-279a33268d9e","Type":"ContainerStarted","Data":"44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.582290 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" event={"ID":"a81cbde5-9395-418f-a205-279a33268d9e","Type":"ContainerStarted","Data":"597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.594550 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.597098 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.597120 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.597130 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.597142 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.597152 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:24Z","lastTransitionTime":"2025-12-06T04:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.607126 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.618988 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.628635 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.644993 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.661695 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.673689 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.690888 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.699384 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.699424 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.699434 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.699451 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.699462 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:24Z","lastTransitionTime":"2025-12-06T04:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.704956 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.718009 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.730208 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.740095 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.754722 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.764048 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.773557 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.788006 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.799549 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.801561 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.801601 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.801616 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.801640 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.801656 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:24Z","lastTransitionTime":"2025-12-06T04:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.819658 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.831675 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.843967 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.855042 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.866861 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.878205 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.888500 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.901596 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.904661 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.904700 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.904711 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.904725 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.904737 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:24Z","lastTransitionTime":"2025-12-06T04:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.917155 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.935794 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.954534 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:24Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.997349 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.997491 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:24 crc kubenswrapper[4718]: E1206 04:07:24.997514 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:07:40.997493267 +0000 UTC m=+50.003198448 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:07:24 crc kubenswrapper[4718]: I1206 04:07:24.997550 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:24 crc kubenswrapper[4718]: E1206 04:07:24.997590 4718 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:24 crc kubenswrapper[4718]: E1206 04:07:24.997621 4718 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:24 crc kubenswrapper[4718]: E1206 04:07:24.997630 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:40.997620749 +0000 UTC m=+50.003325910 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:24 crc kubenswrapper[4718]: E1206 04:07:24.997698 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:40.9976801 +0000 UTC m=+50.003385281 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.007997 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.008040 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.008050 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.008066 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.008077 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.008538 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.026585 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.037672 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-mhz2w"] Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.038269 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.038355 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.050038 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.064220 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.074430 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.086170 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.098110 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.098168 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.098284 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.098301 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.098310 4718 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.098343 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.098362 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:41.098346307 +0000 UTC m=+50.104051468 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.098371 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.098384 4718 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.098425 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 04:07:41.098410878 +0000 UTC m=+50.104116039 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.098352 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.109792 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.109829 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.109842 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.109859 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.109870 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.110158 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.124163 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.134282 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.149206 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.158956 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.168607 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.186370 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.199139 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.199199 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5jlm\" (UniqueName: \"kubernetes.io/projected/38892dbb-af9b-4483-bfe6-c1ed160d356c-kube-api-access-r5jlm\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.205212 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.212866 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.212908 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.212921 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.212938 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.212950 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.222459 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.236494 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.265386 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:25Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.299921 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.299965 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5jlm\" (UniqueName: \"kubernetes.io/projected/38892dbb-af9b-4483-bfe6-c1ed160d356c-kube-api-access-r5jlm\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.300132 4718 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.300202 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs podName:38892dbb-af9b-4483-bfe6-c1ed160d356c nodeName:}" failed. No retries permitted until 2025-12-06 04:07:25.800183919 +0000 UTC m=+34.805889090 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs") pod "network-metrics-daemon-mhz2w" (UID: "38892dbb-af9b-4483-bfe6-c1ed160d356c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.315888 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.315948 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.315966 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.315989 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.316012 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.325871 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5jlm\" (UniqueName: \"kubernetes.io/projected/38892dbb-af9b-4483-bfe6-c1ed160d356c-kube-api-access-r5jlm\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.327437 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.327529 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.327562 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.327583 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.327683 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.327792 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.419777 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.419818 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.419829 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.419852 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.419864 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.522603 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.522660 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.522678 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.522699 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.522715 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.625800 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.625868 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.625886 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.625912 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.625931 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.729457 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.729528 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.729546 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.729572 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.729591 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.804601 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.804792 4718 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:25 crc kubenswrapper[4718]: E1206 04:07:25.804853 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs podName:38892dbb-af9b-4483-bfe6-c1ed160d356c nodeName:}" failed. No retries permitted until 2025-12-06 04:07:26.804836467 +0000 UTC m=+35.810541638 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs") pod "network-metrics-daemon-mhz2w" (UID: "38892dbb-af9b-4483-bfe6-c1ed160d356c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.832528 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.832578 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.832589 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.832609 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.832623 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.935563 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.935637 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.935663 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.935694 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:25 crc kubenswrapper[4718]: I1206 04:07:25.935719 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:25Z","lastTransitionTime":"2025-12-06T04:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.038779 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.038857 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.038880 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.038910 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.038959 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.141596 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.141659 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.141682 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.141702 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.141717 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.244933 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.244980 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.244992 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.245010 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.245022 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.327400 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:26 crc kubenswrapper[4718]: E1206 04:07:26.327543 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.348620 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.348682 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.348699 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.348722 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.348738 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.451705 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.451765 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.451786 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.451813 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.451830 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.553976 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.554020 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.554037 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.554061 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.554079 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.657345 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.657382 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.657395 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.657412 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.657424 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.760546 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.760624 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.760645 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.760674 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.760692 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.818309 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:26 crc kubenswrapper[4718]: E1206 04:07:26.818520 4718 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:26 crc kubenswrapper[4718]: E1206 04:07:26.818645 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs podName:38892dbb-af9b-4483-bfe6-c1ed160d356c nodeName:}" failed. No retries permitted until 2025-12-06 04:07:28.818616801 +0000 UTC m=+37.824322002 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs") pod "network-metrics-daemon-mhz2w" (UID: "38892dbb-af9b-4483-bfe6-c1ed160d356c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.863571 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.863630 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.863644 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.863662 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.863673 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.966747 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.966805 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.966821 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.966846 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:26 crc kubenswrapper[4718]: I1206 04:07:26.966866 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:26Z","lastTransitionTime":"2025-12-06T04:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.068855 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.068898 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.068909 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.069672 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.069685 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.172644 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.172721 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.172746 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.172777 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.172799 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.275017 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.275045 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.275052 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.275064 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.275072 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.327386 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.327521 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:27 crc kubenswrapper[4718]: E1206 04:07:27.327639 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.327696 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:27 crc kubenswrapper[4718]: E1206 04:07:27.327790 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:27 crc kubenswrapper[4718]: E1206 04:07:27.327984 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.378732 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.378809 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.378829 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.378878 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.378896 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.482151 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.482227 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.482283 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.482313 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.482333 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.585155 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.585209 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.585227 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.585310 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.585329 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.688619 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.688687 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.688705 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.688729 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.688752 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.791227 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.791325 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.791346 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.791374 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.791392 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.894129 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.894525 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.894712 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.894907 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.895122 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.998344 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.998390 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.998399 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.998416 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:27 crc kubenswrapper[4718]: I1206 04:07:27.998426 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:27Z","lastTransitionTime":"2025-12-06T04:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.100730 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.100760 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.100770 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.100785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.100795 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:28Z","lastTransitionTime":"2025-12-06T04:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.204658 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.204752 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.204775 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.204808 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.204830 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:28Z","lastTransitionTime":"2025-12-06T04:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.307616 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.307656 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.307668 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.307685 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.307696 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:28Z","lastTransitionTime":"2025-12-06T04:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.327562 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:28 crc kubenswrapper[4718]: E1206 04:07:28.327745 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.410364 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.410403 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.410414 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.410431 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.410471 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:28Z","lastTransitionTime":"2025-12-06T04:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.513093 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.513144 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.513155 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.513171 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.513183 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:28Z","lastTransitionTime":"2025-12-06T04:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.615469 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.615531 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.615543 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.615562 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.615575 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:28Z","lastTransitionTime":"2025-12-06T04:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.718745 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.718785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.718796 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.718811 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.718821 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:28Z","lastTransitionTime":"2025-12-06T04:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.821853 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.821901 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.821914 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.821932 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.821944 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:28Z","lastTransitionTime":"2025-12-06T04:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.841862 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:28 crc kubenswrapper[4718]: E1206 04:07:28.842020 4718 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:28 crc kubenswrapper[4718]: E1206 04:07:28.842077 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs podName:38892dbb-af9b-4483-bfe6-c1ed160d356c nodeName:}" failed. No retries permitted until 2025-12-06 04:07:32.842062067 +0000 UTC m=+41.847767228 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs") pod "network-metrics-daemon-mhz2w" (UID: "38892dbb-af9b-4483-bfe6-c1ed160d356c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.924055 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.924115 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.924132 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.924157 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:28 crc kubenswrapper[4718]: I1206 04:07:28.924174 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:28Z","lastTransitionTime":"2025-12-06T04:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.026842 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.026914 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.026935 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.026964 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.026986 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.129611 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.129656 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.129682 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.129699 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.129710 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.231683 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.231738 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.231756 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.231791 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.231808 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.328137 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.328194 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:29 crc kubenswrapper[4718]: E1206 04:07:29.328307 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.328138 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:29 crc kubenswrapper[4718]: E1206 04:07:29.328481 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:29 crc kubenswrapper[4718]: E1206 04:07:29.328605 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.334570 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.334609 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.334633 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.334656 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.334672 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.437680 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.437752 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.437771 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.437796 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.437814 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.540403 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.540454 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.540469 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.540488 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.540502 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.642500 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.642572 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.642589 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.642613 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.642630 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.745375 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.745418 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.745433 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.745452 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.745469 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.848477 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.848543 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.848562 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.848589 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.848608 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.950969 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.951007 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.951016 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.951028 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:29 crc kubenswrapper[4718]: I1206 04:07:29.951038 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:29Z","lastTransitionTime":"2025-12-06T04:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.053291 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.053352 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.053377 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.053407 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.053431 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.156494 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.156556 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.156572 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.156595 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.156612 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.258659 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.258724 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.258746 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.258775 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.258797 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.327501 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:30 crc kubenswrapper[4718]: E1206 04:07:30.327653 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.360668 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.360735 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.360756 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.360783 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.360804 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.463561 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.463633 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.463660 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.463684 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.463702 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.567018 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.567066 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.567083 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.567109 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.567129 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.670189 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.670318 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.670370 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.670402 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.670419 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.773546 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.773602 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.773625 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.773655 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.773676 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.877695 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.877767 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.877790 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.877818 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.877840 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.981002 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.981061 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.981078 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.981102 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:30 crc kubenswrapper[4718]: I1206 04:07:30.981120 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:30Z","lastTransitionTime":"2025-12-06T04:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.083552 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.083634 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.083657 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.083688 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.083710 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:31Z","lastTransitionTime":"2025-12-06T04:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.186820 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.186884 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.186908 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.186937 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.186959 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:31Z","lastTransitionTime":"2025-12-06T04:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.289095 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.289209 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.289257 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.289281 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.289298 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:31Z","lastTransitionTime":"2025-12-06T04:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.327425 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.327542 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:31 crc kubenswrapper[4718]: E1206 04:07:31.327763 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.327823 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:31 crc kubenswrapper[4718]: E1206 04:07:31.327931 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:31 crc kubenswrapper[4718]: E1206 04:07:31.328039 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.341517 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.354583 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.369489 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.392623 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.392673 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.392720 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.392741 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.392756 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:31Z","lastTransitionTime":"2025-12-06T04:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.403363 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.427117 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.439677 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.453189 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.471021 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.489933 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.495215 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.495352 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.495370 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.495398 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.495415 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:31Z","lastTransitionTime":"2025-12-06T04:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.511198 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.529818 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.552161 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.568040 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.599153 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.599187 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.599200 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.599215 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.599225 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:31Z","lastTransitionTime":"2025-12-06T04:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.601763 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.619013 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.631600 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:31Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.701660 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.701706 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.701722 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.701746 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.701762 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:31Z","lastTransitionTime":"2025-12-06T04:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.804676 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.804738 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.804759 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.804789 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.804809 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:31Z","lastTransitionTime":"2025-12-06T04:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.908311 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.908421 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.908483 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.908513 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:31 crc kubenswrapper[4718]: I1206 04:07:31.908531 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:31Z","lastTransitionTime":"2025-12-06T04:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.010953 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.011009 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.011026 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.011051 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.011069 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.113778 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.114124 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.114142 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.114166 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.114184 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.216910 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.216976 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.216992 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.217018 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.217037 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.320104 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.320161 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.320178 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.320201 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.320220 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.327467 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:32 crc kubenswrapper[4718]: E1206 04:07:32.327741 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.422918 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.422994 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.423006 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.423024 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.423035 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.526118 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.526161 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.526176 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.526196 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.526207 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.628890 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.628967 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.628978 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.629000 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.629013 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.732067 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.732133 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.732150 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.732178 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.732196 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.835797 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.835890 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.835912 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.835939 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.835962 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.884746 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:32 crc kubenswrapper[4718]: E1206 04:07:32.885017 4718 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:32 crc kubenswrapper[4718]: E1206 04:07:32.885176 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs podName:38892dbb-af9b-4483-bfe6-c1ed160d356c nodeName:}" failed. No retries permitted until 2025-12-06 04:07:40.885130853 +0000 UTC m=+49.890836114 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs") pod "network-metrics-daemon-mhz2w" (UID: "38892dbb-af9b-4483-bfe6-c1ed160d356c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.939559 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.939645 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.939669 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.939695 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.939712 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.959015 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.959084 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.959102 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.959131 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.959149 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:32 crc kubenswrapper[4718]: E1206 04:07:32.982074 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:32Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.987871 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.987931 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.987949 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.987975 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:32 crc kubenswrapper[4718]: I1206 04:07:32.987996 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:32Z","lastTransitionTime":"2025-12-06T04:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: E1206 04:07:33.009411 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:33Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.015020 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.015089 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.015112 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.015139 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.015164 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: E1206 04:07:33.034866 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:33Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.039807 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.039866 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.039888 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.039916 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.039939 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: E1206 04:07:33.059032 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:33Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.063852 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.063916 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.063934 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.063959 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.063977 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: E1206 04:07:33.083900 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:33Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:33 crc kubenswrapper[4718]: E1206 04:07:33.084137 4718 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.086025 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.086077 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.086090 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.086109 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.086127 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.188916 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.188965 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.188980 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.189002 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.189020 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.291340 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.291482 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.291508 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.291639 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.291922 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.327929 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.327979 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.328102 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:33 crc kubenswrapper[4718]: E1206 04:07:33.328331 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:33 crc kubenswrapper[4718]: E1206 04:07:33.328502 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:33 crc kubenswrapper[4718]: E1206 04:07:33.328630 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.395718 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.395772 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.395790 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.395811 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.395828 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.498176 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.498307 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.498333 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.498359 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.498427 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.601300 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.601352 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.601368 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.601389 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.601404 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.704338 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.704405 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.704435 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.704461 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.704479 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.807298 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.807370 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.807394 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.807423 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.807446 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.909395 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.909445 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.909456 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.909474 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:33 crc kubenswrapper[4718]: I1206 04:07:33.909488 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:33Z","lastTransitionTime":"2025-12-06T04:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.012412 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.012476 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.012491 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.012515 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.012532 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.115274 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.115348 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.115369 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.115399 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.115418 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.218549 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.218614 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.218635 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.218664 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.218687 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.321579 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.321665 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.321690 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.321722 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.321745 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.327824 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:34 crc kubenswrapper[4718]: E1206 04:07:34.327955 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.423870 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.423945 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.423961 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.423985 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.424003 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.526288 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.526368 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.526387 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.526415 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.526435 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.629032 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.629096 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.629116 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.629144 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.629163 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.732685 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.732726 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.732739 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.732758 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.732771 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.835385 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.835429 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.835440 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.835458 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.835470 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.938744 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.938800 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.938810 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.938834 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:34 crc kubenswrapper[4718]: I1206 04:07:34.938858 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:34Z","lastTransitionTime":"2025-12-06T04:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.041615 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.041692 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.041705 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.041733 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.041749 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.144974 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.145035 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.145053 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.145079 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.145097 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.247546 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.247599 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.247611 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.247630 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.247641 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.328103 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.328162 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:35 crc kubenswrapper[4718]: E1206 04:07:35.328342 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.328489 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:35 crc kubenswrapper[4718]: E1206 04:07:35.328610 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:35 crc kubenswrapper[4718]: E1206 04:07:35.328765 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.350798 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.350841 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.350856 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.350875 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.350895 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.454122 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.454190 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.454208 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.454264 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.454284 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.557435 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.557480 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.557496 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.557518 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.557535 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.660011 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.660089 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.660107 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.660131 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.660149 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.763397 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.763454 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.763472 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.763496 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.763516 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.866292 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.866375 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.866403 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.866439 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.866474 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.969387 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.969446 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.969454 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.969468 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:35 crc kubenswrapper[4718]: I1206 04:07:35.969477 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:35Z","lastTransitionTime":"2025-12-06T04:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.072518 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.072583 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.072599 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.072623 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.072640 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:36Z","lastTransitionTime":"2025-12-06T04:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.176196 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.176323 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.176347 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.176372 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.176388 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:36Z","lastTransitionTime":"2025-12-06T04:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.279420 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.279470 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.279486 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.279509 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.279527 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:36Z","lastTransitionTime":"2025-12-06T04:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.327639 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:36 crc kubenswrapper[4718]: E1206 04:07:36.327877 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.329133 4718 scope.go:117] "RemoveContainer" containerID="792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.382777 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.382845 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.382868 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.382897 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.382920 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:36Z","lastTransitionTime":"2025-12-06T04:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.485976 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.486049 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.486070 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.486101 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.486124 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:36Z","lastTransitionTime":"2025-12-06T04:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.589376 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.589447 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.589466 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.589489 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.589507 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:36Z","lastTransitionTime":"2025-12-06T04:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.693332 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.693387 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.693403 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.693427 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.693443 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:36Z","lastTransitionTime":"2025-12-06T04:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.796388 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.796452 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.796474 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.796502 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.796524 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:36Z","lastTransitionTime":"2025-12-06T04:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.899430 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.899485 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.899502 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.899527 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:36 crc kubenswrapper[4718]: I1206 04:07:36.899543 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:36Z","lastTransitionTime":"2025-12-06T04:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.002362 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.002432 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.002449 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.002473 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.002490 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.105454 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.105525 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.105542 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.105566 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.105583 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.208650 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.208715 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.208738 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.208765 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.208788 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.311160 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.311216 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.311265 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.311290 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.311305 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.327862 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:37 crc kubenswrapper[4718]: E1206 04:07:37.328167 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.327835 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:37 crc kubenswrapper[4718]: E1206 04:07:37.328593 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.327893 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:37 crc kubenswrapper[4718]: E1206 04:07:37.329000 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.413603 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.413677 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.413699 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.413728 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.413752 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.516745 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.517117 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.517306 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.517466 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.517621 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.620894 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.620952 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.620968 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.620992 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.621011 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.723909 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.723997 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.724023 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.724059 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.724081 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.826876 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.827420 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.827643 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.827852 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.828039 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.930662 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.930721 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.930738 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.930762 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:37 crc kubenswrapper[4718]: I1206 04:07:37.930778 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:37Z","lastTransitionTime":"2025-12-06T04:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.033489 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.033527 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.033538 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.033554 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.033564 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.136702 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.136767 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.136785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.136814 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.136832 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.240469 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.240852 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.241087 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.241278 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.241492 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.327479 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:38 crc kubenswrapper[4718]: E1206 04:07:38.327690 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.344667 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.344733 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.344751 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.344775 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.344794 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.448061 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.448143 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.448156 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.448195 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.448209 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.551406 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.551473 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.551489 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.551513 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.551529 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.649828 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/1.log" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.653431 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.653462 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.653471 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.653483 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.653491 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.653515 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.654199 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.671319 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.682704 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.692975 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.709292 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.719403 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.730311 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.749788 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.755376 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.755438 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.755450 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.755466 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.755476 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.760690 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.771756 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.783727 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.793822 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.807569 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.830197 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.849443 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.858550 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.858598 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.858608 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.858744 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.858762 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.866949 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.878944 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:38Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.961351 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.961451 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.961470 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.961494 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:38 crc kubenswrapper[4718]: I1206 04:07:38.961511 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:38Z","lastTransitionTime":"2025-12-06T04:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.063986 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.064638 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.064705 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.064784 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.064852 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.168308 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.168379 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.168396 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.168421 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.168439 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.271097 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.271287 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.271319 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.271350 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.271373 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.328295 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.328394 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.328417 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:39 crc kubenswrapper[4718]: E1206 04:07:39.328534 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:39 crc kubenswrapper[4718]: E1206 04:07:39.328847 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:39 crc kubenswrapper[4718]: E1206 04:07:39.328705 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.373970 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.374005 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.374014 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.374027 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.374038 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.477939 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.478029 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.478081 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.478105 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.478123 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.580647 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.580953 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.581091 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.581282 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.581441 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.686772 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.686854 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.686868 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.686890 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.686901 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.790360 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.790447 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.790472 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.790504 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.790538 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.894280 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.894323 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.894334 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.894349 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.894359 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.997934 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.997989 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.998005 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.998026 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:39 crc kubenswrapper[4718]: I1206 04:07:39.998043 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:39Z","lastTransitionTime":"2025-12-06T04:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.101404 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.101466 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.101486 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.101510 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.101527 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:40Z","lastTransitionTime":"2025-12-06T04:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.205495 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.205595 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.205615 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.205640 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.205658 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:40Z","lastTransitionTime":"2025-12-06T04:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.308097 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.308156 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.308167 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.308185 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.308196 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:40Z","lastTransitionTime":"2025-12-06T04:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.327725 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:40 crc kubenswrapper[4718]: E1206 04:07:40.327876 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.411550 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.411630 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.411645 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.411671 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.411687 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:40Z","lastTransitionTime":"2025-12-06T04:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.515153 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.515207 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.515225 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.515300 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.515320 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:40Z","lastTransitionTime":"2025-12-06T04:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.618326 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.618390 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.618404 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.618429 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.618445 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:40Z","lastTransitionTime":"2025-12-06T04:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.662107 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/2.log" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.663035 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/1.log" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.666595 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463" exitCode=1 Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.666848 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.666909 4718 scope.go:117] "RemoveContainer" containerID="792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.669075 4718 scope.go:117] "RemoveContainer" containerID="12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463" Dec 06 04:07:40 crc kubenswrapper[4718]: E1206 04:07:40.669443 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.695163 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.707730 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.721473 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.721537 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.721554 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.721580 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.721598 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:40Z","lastTransitionTime":"2025-12-06T04:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.729862 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.746311 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.761860 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.778303 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.802149 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.823359 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.825933 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.825998 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.826017 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.826043 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.826061 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:40Z","lastTransitionTime":"2025-12-06T04:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.842123 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.861072 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.882026 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.902685 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.919143 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.928341 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.928585 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.928675 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.928787 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.928877 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:40Z","lastTransitionTime":"2025-12-06T04:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.935153 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.959342 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.971017 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:40 crc kubenswrapper[4718]: E1206 04:07:40.971283 4718 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:40 crc kubenswrapper[4718]: E1206 04:07:40.971436 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs podName:38892dbb-af9b-4483-bfe6-c1ed160d356c nodeName:}" failed. No retries permitted until 2025-12-06 04:07:56.971404427 +0000 UTC m=+65.977109628 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs") pod "network-metrics-daemon-mhz2w" (UID: "38892dbb-af9b-4483-bfe6-c1ed160d356c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:40 crc kubenswrapper[4718]: I1206 04:07:40.976179 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:40Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.032361 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.032425 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.032442 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.032465 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.032480 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.072959 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.073161 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:08:13.073119824 +0000 UTC m=+82.078824985 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.073306 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.073339 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.073445 4718 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.073506 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:08:13.073489831 +0000 UTC m=+82.079195002 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.073513 4718 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.073560 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:08:13.073549933 +0000 UTC m=+82.079255194 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.135767 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.135839 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.135860 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.135888 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.135913 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.174393 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.174511 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.174679 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.174727 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.174796 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.174821 4718 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.174897 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 04:08:13.174872432 +0000 UTC m=+82.180577623 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.174736 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.174935 4718 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.175010 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 04:08:13.174984444 +0000 UTC m=+82.180689665 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.239181 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.239279 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.239327 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.239357 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.239377 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.328154 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.328356 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.328426 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.328650 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.328895 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:41 crc kubenswrapper[4718]: E1206 04:07:41.329170 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.343819 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.343861 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.343872 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.343887 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.343898 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.345759 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.372422 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.389171 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.403600 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.422681 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.437973 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.446593 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.446892 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.447099 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.447397 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.447595 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.455912 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.470833 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.481013 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.498912 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.515789 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.531910 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.550860 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.550896 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.550906 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.550919 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.550928 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.551192 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.565434 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.570701 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.575852 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.587675 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.603453 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.619009 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.640115 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.653451 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.653475 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.653483 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.653521 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.653531 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.658410 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.670926 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.672326 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/2.log" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.701380 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.711319 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.729115 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.739556 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.755754 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.755841 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.755853 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.755870 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.755881 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.756726 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.769683 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.781266 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.794987 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.805793 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.816699 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.834044 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.846077 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.857121 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:41Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.859008 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.859056 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.859074 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.859098 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.859114 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.961318 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.961388 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.961409 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.961435 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:41 crc kubenswrapper[4718]: I1206 04:07:41.961453 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:41Z","lastTransitionTime":"2025-12-06T04:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.064713 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.064781 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.064798 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.064823 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.064841 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.167961 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.168436 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.168465 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.168499 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.168521 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.271908 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.271973 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.271992 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.272015 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.272032 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.327880 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:42 crc kubenswrapper[4718]: E1206 04:07:42.328076 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.375190 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.375264 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.375279 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.375300 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.375314 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.478442 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.478527 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.478557 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.478592 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.478616 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.580889 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.580945 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.580957 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.580975 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.580987 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.682838 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.682877 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.682890 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.682907 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.682919 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.786857 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.786916 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.786933 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.786957 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.786973 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.889843 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.889900 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.889919 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.889944 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.889962 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.992607 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.992677 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.992697 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.992724 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:42 crc kubenswrapper[4718]: I1206 04:07:42.992743 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:42Z","lastTransitionTime":"2025-12-06T04:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.095764 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.095829 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.095849 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.095876 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.095892 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.122048 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.122107 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.122125 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.122151 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.122168 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: E1206 04:07:43.144630 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:43Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.149562 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.149613 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.149630 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.149653 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.149668 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: E1206 04:07:43.170328 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:43Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.175463 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.175535 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.175553 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.175583 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.175603 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: E1206 04:07:43.195857 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:43Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.200630 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.200668 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.200685 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.200707 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.200725 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: E1206 04:07:43.217649 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:43Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.222118 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.222172 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.222193 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.222218 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.222271 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: E1206 04:07:43.245121 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:43Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:43 crc kubenswrapper[4718]: E1206 04:07:43.245413 4718 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.247904 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.247985 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.248003 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.248026 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.248042 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.328301 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.328363 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.328493 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:43 crc kubenswrapper[4718]: E1206 04:07:43.328714 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:43 crc kubenswrapper[4718]: E1206 04:07:43.328808 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:43 crc kubenswrapper[4718]: E1206 04:07:43.329226 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.350934 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.351003 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.351020 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.351040 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.351482 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.455430 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.455529 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.455546 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.455571 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.455591 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.559039 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.559096 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.559113 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.559141 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.559162 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.662679 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.662725 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.662741 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.662765 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.662784 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.766293 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.766356 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.766379 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.766410 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.766432 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.868997 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.869062 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.869084 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.869112 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.869132 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.972058 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.972147 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.972169 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.972196 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:43 crc kubenswrapper[4718]: I1206 04:07:43.972216 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:43Z","lastTransitionTime":"2025-12-06T04:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.075000 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.075038 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.075049 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.075064 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.075076 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:44Z","lastTransitionTime":"2025-12-06T04:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.177854 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.177947 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.177961 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.177978 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.177988 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:44Z","lastTransitionTime":"2025-12-06T04:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.280329 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.280398 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.280418 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.280444 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.280461 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:44Z","lastTransitionTime":"2025-12-06T04:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.327648 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:44 crc kubenswrapper[4718]: E1206 04:07:44.327849 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.382803 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.382871 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.382889 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.382913 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.382931 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:44Z","lastTransitionTime":"2025-12-06T04:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.485436 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.485489 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.485501 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.485518 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.485530 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:44Z","lastTransitionTime":"2025-12-06T04:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.588827 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.588896 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.588913 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.588938 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.588955 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:44Z","lastTransitionTime":"2025-12-06T04:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.691813 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.691893 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.691918 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.691955 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.691982 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:44Z","lastTransitionTime":"2025-12-06T04:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.795085 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.795167 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.795191 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.795225 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.795279 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:44Z","lastTransitionTime":"2025-12-06T04:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.898068 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.898138 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.898161 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.898189 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:44 crc kubenswrapper[4718]: I1206 04:07:44.898211 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:44Z","lastTransitionTime":"2025-12-06T04:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.001299 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.001370 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.001395 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.001420 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.001437 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.105898 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.106326 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.106518 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.106673 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.106804 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.209837 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.209910 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.209929 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.209954 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.209972 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.312753 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.312810 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.312822 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.312844 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.312860 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.328464 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.328646 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:45 crc kubenswrapper[4718]: E1206 04:07:45.329038 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:45 crc kubenswrapper[4718]: E1206 04:07:45.328825 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.328680 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:45 crc kubenswrapper[4718]: E1206 04:07:45.329195 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.415066 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.415109 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.415140 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.415156 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.415180 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.519626 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.519694 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.519715 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.519740 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.519765 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.623071 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.623121 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.623139 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.623163 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.623179 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.725582 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.725634 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.725656 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.725702 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.725726 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.829778 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.830287 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.830439 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.830605 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.830802 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.933839 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.934199 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.934402 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.934611 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:45 crc kubenswrapper[4718]: I1206 04:07:45.934764 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:45Z","lastTransitionTime":"2025-12-06T04:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.038298 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.038352 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.038367 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.038390 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.038407 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.141089 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.141488 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.141658 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.141804 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.141941 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.243924 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.244536 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.244610 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.244686 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.244758 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.327471 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:46 crc kubenswrapper[4718]: E1206 04:07:46.327683 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.347803 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.347852 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.347870 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.347891 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.347909 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.450532 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.450587 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.450600 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.450619 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.450632 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.553871 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.553938 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.553960 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.553993 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.554016 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.656318 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.656368 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.656387 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.656411 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.656428 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.759563 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.760022 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.760312 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.760581 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.760802 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.863882 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.863943 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.863960 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.863985 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.864003 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.967584 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.967701 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.967720 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.967783 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:46 crc kubenswrapper[4718]: I1206 04:07:46.967800 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:46Z","lastTransitionTime":"2025-12-06T04:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.071511 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.071560 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.071599 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.071617 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.071628 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:47Z","lastTransitionTime":"2025-12-06T04:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.175007 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.175081 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.175099 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.175127 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.175150 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:47Z","lastTransitionTime":"2025-12-06T04:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.279577 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.279629 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.279643 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.279663 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.279678 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:47Z","lastTransitionTime":"2025-12-06T04:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.327994 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.328129 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:47 crc kubenswrapper[4718]: E1206 04:07:47.328208 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.328280 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:47 crc kubenswrapper[4718]: E1206 04:07:47.328531 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:47 crc kubenswrapper[4718]: E1206 04:07:47.328593 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.382898 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.383281 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.383470 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.383581 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.383672 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:47Z","lastTransitionTime":"2025-12-06T04:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.486208 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.486299 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.486317 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.486344 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.486361 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:47Z","lastTransitionTime":"2025-12-06T04:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.589468 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.589914 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.590076 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.590330 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.590585 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:47Z","lastTransitionTime":"2025-12-06T04:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.694069 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.694127 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.694153 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.694182 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.694206 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:47Z","lastTransitionTime":"2025-12-06T04:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.797156 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.797527 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.797729 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.797934 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.798106 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:47Z","lastTransitionTime":"2025-12-06T04:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.901431 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.901855 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.902062 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.902333 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:47 crc kubenswrapper[4718]: I1206 04:07:47.902548 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:47Z","lastTransitionTime":"2025-12-06T04:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.006598 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.006666 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.006682 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.006708 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.006725 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.109785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.110048 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.110090 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.110117 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.110139 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.213585 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.213636 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.213650 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.213672 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.213717 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.316728 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.316792 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.316809 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.316833 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.316850 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.327460 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:48 crc kubenswrapper[4718]: E1206 04:07:48.327665 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.419619 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.419667 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.419680 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.419701 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.419712 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.523788 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.524535 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.524626 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.524662 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.524683 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.627689 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.627757 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.627778 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.627807 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.627828 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.730906 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.730981 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.730991 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.731011 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.731023 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.834620 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.834692 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.834715 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.834743 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.834760 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.937952 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.938293 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.938408 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.938528 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:48 crc kubenswrapper[4718]: I1206 04:07:48.938671 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:48Z","lastTransitionTime":"2025-12-06T04:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.041825 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.041892 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.041908 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.041932 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.041950 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.145186 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.145259 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.145271 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.145291 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.145305 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.248818 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.248912 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.248939 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.248982 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.249012 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.327883 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.327992 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.328023 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:49 crc kubenswrapper[4718]: E1206 04:07:49.328740 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:49 crc kubenswrapper[4718]: E1206 04:07:49.328791 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:49 crc kubenswrapper[4718]: E1206 04:07:49.329056 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.352592 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.352673 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.352701 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.352731 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.352754 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.455899 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.455947 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.455961 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.455981 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.455997 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.558304 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.558359 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.558369 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.558384 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.558397 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.661725 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.661853 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.661880 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.661909 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.661928 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.764618 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.764713 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.764732 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.764759 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.764777 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.867911 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.867957 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.867977 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.867999 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.868014 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.970968 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.971037 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.971055 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.971083 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:49 crc kubenswrapper[4718]: I1206 04:07:49.971103 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:49Z","lastTransitionTime":"2025-12-06T04:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.074369 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.074413 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.074425 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.074446 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.074459 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:50Z","lastTransitionTime":"2025-12-06T04:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.177729 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.177794 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.177807 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.177834 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.177850 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:50Z","lastTransitionTime":"2025-12-06T04:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.281453 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.281509 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.281524 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.281547 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.281562 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:50Z","lastTransitionTime":"2025-12-06T04:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.327311 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:50 crc kubenswrapper[4718]: E1206 04:07:50.327563 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.389678 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.389781 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.389810 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.389850 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.389873 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:50Z","lastTransitionTime":"2025-12-06T04:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.492344 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.492417 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.492441 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.492472 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.492496 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:50Z","lastTransitionTime":"2025-12-06T04:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.595889 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.596009 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.596035 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.596067 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.596091 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:50Z","lastTransitionTime":"2025-12-06T04:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.699826 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.699915 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.699950 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.699984 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.700008 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:50Z","lastTransitionTime":"2025-12-06T04:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.803613 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.803690 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.803708 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.803734 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.803753 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:50Z","lastTransitionTime":"2025-12-06T04:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.907097 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.907168 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.907193 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.907222 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:50 crc kubenswrapper[4718]: I1206 04:07:50.907360 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:50Z","lastTransitionTime":"2025-12-06T04:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.010361 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.010440 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.010465 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.010494 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.010518 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.113991 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.114043 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.114057 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.114077 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.114090 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.217302 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.217344 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.217356 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.217372 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.217383 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.319684 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.319726 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.319738 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.319754 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.319765 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.327334 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.327365 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.327345 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:51 crc kubenswrapper[4718]: E1206 04:07:51.327498 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:51 crc kubenswrapper[4718]: E1206 04:07:51.327563 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:51 crc kubenswrapper[4718]: E1206 04:07:51.327631 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.328990 4718 scope.go:117] "RemoveContainer" containerID="12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463" Dec 06 04:07:51 crc kubenswrapper[4718]: E1206 04:07:51.329365 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.346195 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.359744 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.371371 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.383783 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.400841 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.417616 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.421572 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.421636 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.421652 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.421674 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.421692 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.433481 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.464970 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://792d6f21321b56b29100b340c523b1144670761fe635993ec9ec24eb901de426\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:22Z\\\",\\\"message\\\":\\\"flector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:22.396491 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:22.396984 6129 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:22.397043 6129 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 04:07:22.397052 6129 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 04:07:22.397097 6129 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:22.397111 6129 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:22.397128 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 04:07:22.397123 6129 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:22.397141 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 04:07:22.397155 6129 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:22.397175 6129 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:22.397202 6129 factory.go:656] Stopping watch factory\\\\nI1206 04:07:22.397220 6129 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 04:07:22.397228 6129 ovnkube.go:599] Stopped ovnkube\\\\nI1206 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.482584 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.501211 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.518910 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.525481 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.525565 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.525585 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.525617 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.525640 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.538115 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.557090 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.575644 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.592142 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.611437 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.624541 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.628600 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.628666 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.628686 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.628711 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.628724 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.638294 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.653989 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.685359 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.702614 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.717912 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.730754 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.730795 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.730804 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.730820 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.730831 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.733811 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.747735 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.761497 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.776474 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.788893 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.804805 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.827335 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.833792 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.833834 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.833844 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.833862 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.833874 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.841901 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.862391 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.875424 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.895138 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.909932 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:51Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.937540 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.937899 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.938037 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.938111 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:51 crc kubenswrapper[4718]: I1206 04:07:51.938198 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:51Z","lastTransitionTime":"2025-12-06T04:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.042412 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.042458 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.042474 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.042497 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.042513 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.145022 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.145078 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.145096 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.145124 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.145142 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.249025 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.249095 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.249111 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.249137 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.249155 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.327490 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:52 crc kubenswrapper[4718]: E1206 04:07:52.328116 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.352136 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.352199 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.352219 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.352281 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.352301 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.454906 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.454965 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.454977 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.454997 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.455008 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.557608 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.557656 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.557669 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.557686 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.557695 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.660328 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.660400 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.660419 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.660446 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.660464 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.762587 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.762645 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.762666 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.762693 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.762713 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.865572 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.865628 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.865645 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.865668 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.865687 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.968872 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.968940 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.968959 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.968984 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:52 crc kubenswrapper[4718]: I1206 04:07:52.969002 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:52Z","lastTransitionTime":"2025-12-06T04:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.072101 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.072181 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.072204 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.072275 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.072300 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.174903 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.174936 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.174944 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.174957 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.174966 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.278300 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.278361 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.278380 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.278403 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.278421 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.285400 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.285489 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.285506 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.285530 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.285547 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: E1206 04:07:53.303825 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:53Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.307696 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.307741 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.307753 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.307771 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.307782 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: E1206 04:07:53.322138 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:53Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.325811 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.325847 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.325858 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.325875 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.325887 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.327715 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.327791 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:53 crc kubenswrapper[4718]: E1206 04:07:53.327878 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.327960 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:53 crc kubenswrapper[4718]: E1206 04:07:53.328059 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:53 crc kubenswrapper[4718]: E1206 04:07:53.328186 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:53 crc kubenswrapper[4718]: E1206 04:07:53.342056 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:53Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.346169 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.346218 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.346248 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.346272 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.346287 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: E1206 04:07:53.362949 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:53Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.370841 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.370884 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.370896 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.370915 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.370932 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: E1206 04:07:53.388342 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:53Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:53 crc kubenswrapper[4718]: E1206 04:07:53.388834 4718 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.390836 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.390916 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.390929 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.390955 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.390969 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.511623 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.511692 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.511710 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.511737 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.511753 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.613932 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.613987 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.614004 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.614047 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.614063 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.717190 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.717262 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.717276 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.717292 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.717303 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.820673 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.820746 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.820768 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.820811 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.820848 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.923965 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.924013 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.924030 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.924051 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:53 crc kubenswrapper[4718]: I1206 04:07:53.924067 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:53Z","lastTransitionTime":"2025-12-06T04:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.026907 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.026981 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.027000 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.027032 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.027050 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.130502 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.130599 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.130638 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.130670 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.130693 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.233935 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.233996 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.234014 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.234042 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.234064 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.327852 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:54 crc kubenswrapper[4718]: E1206 04:07:54.328044 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.335899 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.335964 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.335984 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.336010 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.336030 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.438715 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.438780 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.438802 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.438831 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.438853 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.541426 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.541482 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.541499 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.541522 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.541538 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.644576 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.644633 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.644650 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.644681 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.644699 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.746972 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.747013 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.747025 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.747042 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.747054 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.850515 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.850594 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.850614 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.850640 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.850657 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.954193 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.954298 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.954318 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.954344 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:54 crc kubenswrapper[4718]: I1206 04:07:54.954363 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:54Z","lastTransitionTime":"2025-12-06T04:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.056309 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.056345 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.056355 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.056369 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.056380 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.158801 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.158857 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.158875 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.158899 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.158914 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.262514 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.262578 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.262594 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.262619 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.262638 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.327595 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:55 crc kubenswrapper[4718]: E1206 04:07:55.327721 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.327888 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:55 crc kubenswrapper[4718]: E1206 04:07:55.327934 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.328123 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:55 crc kubenswrapper[4718]: E1206 04:07:55.328190 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.365344 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.365392 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.365404 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.365422 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.365435 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.468886 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.468928 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.468937 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.468951 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.468961 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.571846 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.571896 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.571912 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.571928 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.571939 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.674382 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.674452 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.674471 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.674496 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.674514 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.776765 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.776811 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.776822 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.776838 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.776849 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.879078 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.879131 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.879141 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.879160 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.879170 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.981142 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.981183 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.981193 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.981213 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:55 crc kubenswrapper[4718]: I1206 04:07:55.981224 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:55Z","lastTransitionTime":"2025-12-06T04:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.083581 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.083618 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.083627 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.083641 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.083650 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:56Z","lastTransitionTime":"2025-12-06T04:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.186088 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.186129 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.186139 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.186155 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.186166 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:56Z","lastTransitionTime":"2025-12-06T04:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.288786 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.288832 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.288842 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.288857 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.288869 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:56Z","lastTransitionTime":"2025-12-06T04:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.327922 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:56 crc kubenswrapper[4718]: E1206 04:07:56.328042 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.391049 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.391100 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.391117 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.391143 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.391170 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:56Z","lastTransitionTime":"2025-12-06T04:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.493787 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.493860 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.493879 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.493904 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.493922 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:56Z","lastTransitionTime":"2025-12-06T04:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.596484 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.596542 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.596561 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.596672 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.596692 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:56Z","lastTransitionTime":"2025-12-06T04:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.700087 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.700148 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.700166 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.700189 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.700205 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:56Z","lastTransitionTime":"2025-12-06T04:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.802749 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.802882 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.802943 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.802966 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.802984 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:56Z","lastTransitionTime":"2025-12-06T04:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.907975 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.908044 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.908055 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.908070 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:56 crc kubenswrapper[4718]: I1206 04:07:56.908080 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:56Z","lastTransitionTime":"2025-12-06T04:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.011132 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.011203 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.011216 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.011262 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.011281 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.059418 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:57 crc kubenswrapper[4718]: E1206 04:07:57.059674 4718 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:57 crc kubenswrapper[4718]: E1206 04:07:57.059839 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs podName:38892dbb-af9b-4483-bfe6-c1ed160d356c nodeName:}" failed. No retries permitted until 2025-12-06 04:08:29.059803818 +0000 UTC m=+98.065509199 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs") pod "network-metrics-daemon-mhz2w" (UID: "38892dbb-af9b-4483-bfe6-c1ed160d356c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.113397 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.113466 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.113480 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.113502 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.113516 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.216996 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.217068 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.217083 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.217109 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.217130 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.320218 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.320317 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.320337 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.320365 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.320384 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.328510 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:57 crc kubenswrapper[4718]: E1206 04:07:57.328701 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.328996 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:57 crc kubenswrapper[4718]: E1206 04:07:57.329095 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.329463 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:57 crc kubenswrapper[4718]: E1206 04:07:57.329567 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.423381 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.423444 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.423464 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.423491 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.423508 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.527188 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.527280 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.527297 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.527321 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.527338 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.630782 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.630821 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.630831 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.630845 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.630855 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.732434 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.732476 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.732487 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.732503 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.732546 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.834950 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.834992 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.835005 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.835020 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.835028 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.937494 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.937537 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.937547 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.937561 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:57 crc kubenswrapper[4718]: I1206 04:07:57.937570 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:57Z","lastTransitionTime":"2025-12-06T04:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.041807 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.041880 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.041892 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.041915 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.041928 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.144282 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.144331 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.144347 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.144362 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.144372 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.246984 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.247030 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.247040 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.247058 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.247070 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.328008 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:07:58 crc kubenswrapper[4718]: E1206 04:07:58.328242 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.349297 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.349416 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.349440 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.349520 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.349545 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.452349 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.452397 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.452409 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.452426 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.452439 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.554990 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.555038 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.555049 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.555064 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.555075 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.658732 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.658768 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.659145 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.659166 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.659179 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.774107 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.774141 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.774149 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.774161 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.774172 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.876684 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.876732 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.876744 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.876761 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.876773 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.979109 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.979159 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.979171 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.979188 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:58 crc kubenswrapper[4718]: I1206 04:07:58.979200 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:58Z","lastTransitionTime":"2025-12-06T04:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.081788 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.081828 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.081837 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.081852 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.081861 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:59Z","lastTransitionTime":"2025-12-06T04:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.184700 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.184753 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.184764 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.184779 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.184788 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:59Z","lastTransitionTime":"2025-12-06T04:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.288514 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.288554 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.288565 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.288579 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.288589 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:59Z","lastTransitionTime":"2025-12-06T04:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.327611 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.327710 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.327633 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:07:59 crc kubenswrapper[4718]: E1206 04:07:59.327797 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:07:59 crc kubenswrapper[4718]: E1206 04:07:59.327960 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:07:59 crc kubenswrapper[4718]: E1206 04:07:59.328094 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.391487 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.391520 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.391531 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.391546 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.391557 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:59Z","lastTransitionTime":"2025-12-06T04:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.494407 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.494479 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.494511 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.494540 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.494558 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:59Z","lastTransitionTime":"2025-12-06T04:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.596629 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.596671 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.596686 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.596703 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.596715 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:59Z","lastTransitionTime":"2025-12-06T04:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.699286 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.699330 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.699339 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.699357 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.699366 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:59Z","lastTransitionTime":"2025-12-06T04:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.740382 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4r7xc_2272fbce-ce87-4f9f-b1c5-f83f262534de/kube-multus/0.log" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.740430 4718 generic.go:334] "Generic (PLEG): container finished" podID="2272fbce-ce87-4f9f-b1c5-f83f262534de" containerID="7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621" exitCode=1 Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.740460 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4r7xc" event={"ID":"2272fbce-ce87-4f9f-b1c5-f83f262534de","Type":"ContainerDied","Data":"7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.740824 4718 scope.go:117] "RemoveContainer" containerID="7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.757090 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.778924 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.797182 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.801954 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.802010 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.802023 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.802043 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.802057 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:59Z","lastTransitionTime":"2025-12-06T04:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.813263 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:59Z\\\",\\\"message\\\":\\\"2025-12-06T04:07:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389\\\\n2025-12-06T04:07:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389 to /host/opt/cni/bin/\\\\n2025-12-06T04:07:14Z [verbose] multus-daemon started\\\\n2025-12-06T04:07:14Z [verbose] Readiness Indicator file check\\\\n2025-12-06T04:07:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.825921 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.841968 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.854804 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.865825 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.875126 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.889659 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.902005 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.904095 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.904130 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.904142 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.904162 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.904172 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:07:59Z","lastTransitionTime":"2025-12-06T04:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.917988 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.945564 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.965297 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.981484 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:07:59 crc kubenswrapper[4718]: I1206 04:07:59.993805 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:07:59Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.006863 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.006894 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.006903 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.006918 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.006929 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.007264 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.109345 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.109377 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.109385 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.109409 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.109419 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.212496 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.212555 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.212572 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.212608 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.212643 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.314864 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.314906 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.314915 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.314933 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.314946 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.327348 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:00 crc kubenswrapper[4718]: E1206 04:08:00.327566 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.417998 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.418045 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.418059 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.418080 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.418094 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.520418 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.520490 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.520505 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.520529 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.520543 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.623934 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.624002 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.624012 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.624035 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.624049 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.727171 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.727213 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.727225 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.727249 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.727302 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.745509 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4r7xc_2272fbce-ce87-4f9f-b1c5-f83f262534de/kube-multus/0.log" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.745574 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4r7xc" event={"ID":"2272fbce-ce87-4f9f-b1c5-f83f262534de","Type":"ContainerStarted","Data":"c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.763424 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.773174 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.800331 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.816132 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.830014 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.830057 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.830068 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.830088 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.830102 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.830853 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.843660 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.858961 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.869925 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.886311 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.910067 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.927842 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:59Z\\\",\\\"message\\\":\\\"2025-12-06T04:07:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389\\\\n2025-12-06T04:07:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389 to /host/opt/cni/bin/\\\\n2025-12-06T04:07:14Z [verbose] multus-daemon started\\\\n2025-12-06T04:07:14Z [verbose] Readiness Indicator file check\\\\n2025-12-06T04:07:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.932369 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.932427 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.932446 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.932469 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.932485 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:00Z","lastTransitionTime":"2025-12-06T04:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.939120 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.951372 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.961876 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:00 crc kubenswrapper[4718]: I1206 04:08:00.976723 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:00Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.004195 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.014611 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.035204 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.035243 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.035269 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.035286 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.035294 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.137849 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.137905 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.137922 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.137945 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.137962 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.240711 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.240744 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.240753 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.240765 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.240774 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.328255 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.328308 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:01 crc kubenswrapper[4718]: E1206 04:08:01.328839 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.328854 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:01 crc kubenswrapper[4718]: E1206 04:08:01.328934 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:01 crc kubenswrapper[4718]: E1206 04:08:01.329016 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.341120 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.342374 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.342505 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.342632 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.342748 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.343962 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.343840 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.354934 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.365132 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.373908 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.401302 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.414968 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.446536 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.448431 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.448466 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.448478 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.448494 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.448507 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.463906 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.484072 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.497924 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.510344 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.522156 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.534906 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:59Z\\\",\\\"message\\\":\\\"2025-12-06T04:07:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389\\\\n2025-12-06T04:07:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389 to /host/opt/cni/bin/\\\\n2025-12-06T04:07:14Z [verbose] multus-daemon started\\\\n2025-12-06T04:07:14Z [verbose] Readiness Indicator file check\\\\n2025-12-06T04:07:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.546440 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.550515 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.550555 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.550569 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.550587 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.550600 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.560449 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.572448 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.582799 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:01Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.652729 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.652774 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.652788 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.652806 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.652833 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.754318 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.754369 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.754381 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.754397 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.754409 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.857802 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.857871 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.857888 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.857909 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.857925 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.961314 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.961359 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.961374 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.961395 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:01 crc kubenswrapper[4718]: I1206 04:08:01.961409 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:01Z","lastTransitionTime":"2025-12-06T04:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.064159 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.064210 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.064223 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.064267 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.064279 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.167299 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.167371 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.167393 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.167423 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.167440 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.271402 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.271463 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.271480 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.271503 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.271522 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.327725 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:02 crc kubenswrapper[4718]: E1206 04:08:02.327976 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.374190 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.374292 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.374311 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.374338 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.374357 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.477583 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.477654 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.477676 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.477705 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.477727 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.580915 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.581105 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.581119 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.581134 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.581147 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.683232 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.683305 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.683325 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.683341 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.683353 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.785491 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.785545 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.785561 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.785583 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.785601 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.888453 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.888503 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.888520 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.888542 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.888559 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.991603 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.991665 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.991681 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.991704 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:02 crc kubenswrapper[4718]: I1206 04:08:02.991722 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:02Z","lastTransitionTime":"2025-12-06T04:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.094269 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.094331 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.094352 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.094382 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.094407 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.197165 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.197219 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.197287 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.197333 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.197355 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.300193 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.300285 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.300303 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.300328 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.300345 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.327275 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.327309 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.327376 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:03 crc kubenswrapper[4718]: E1206 04:08:03.327423 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:03 crc kubenswrapper[4718]: E1206 04:08:03.327586 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:03 crc kubenswrapper[4718]: E1206 04:08:03.327662 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.403049 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.403129 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.403163 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.403195 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.403216 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.475465 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.475538 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.475556 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.475581 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.475600 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: E1206 04:08:03.491934 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:03Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.497905 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.497969 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.497987 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.498010 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.498029 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: E1206 04:08:03.523947 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:03Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.529332 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.529499 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.529528 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.530176 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.530475 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: E1206 04:08:03.544955 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:03Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.549094 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.549163 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.549174 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.549212 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.549226 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: E1206 04:08:03.566972 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:03Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.572206 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.572265 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.572276 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.572289 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.572298 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: E1206 04:08:03.592372 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:03Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:03 crc kubenswrapper[4718]: E1206 04:08:03.592773 4718 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.594552 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.594612 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.594629 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.594653 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.594671 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.696549 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.696597 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.696606 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.696623 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.696633 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.799538 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.799588 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.799606 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.799631 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.799652 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.902623 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.902761 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.902786 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.902813 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:03 crc kubenswrapper[4718]: I1206 04:08:03.902840 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:03Z","lastTransitionTime":"2025-12-06T04:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.004732 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.004765 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.004775 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.004788 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.004798 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.106870 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.106921 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.106934 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.106952 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.106965 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.209189 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.209268 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.209287 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.209311 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.209327 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.312353 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.312585 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.312607 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.312638 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.312658 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.328111 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:04 crc kubenswrapper[4718]: E1206 04:08:04.328371 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.416302 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.416368 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.416386 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.416412 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.416432 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.520109 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.520164 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.520181 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.520201 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.520215 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.622631 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.622675 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.622686 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.622702 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.622715 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.725379 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.725408 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.725436 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.725450 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.725459 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.828722 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.828801 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.828816 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.828836 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.828848 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.932101 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.932152 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.932163 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.932191 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:04 crc kubenswrapper[4718]: I1206 04:08:04.932206 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:04Z","lastTransitionTime":"2025-12-06T04:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.035100 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.035178 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.035195 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.035222 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.035324 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.138076 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.138134 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.138143 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.138157 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.138168 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.241541 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.241660 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.241680 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.241707 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.241724 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.327852 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.327901 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:05 crc kubenswrapper[4718]: E1206 04:08:05.328049 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.328138 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:05 crc kubenswrapper[4718]: E1206 04:08:05.328406 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:05 crc kubenswrapper[4718]: E1206 04:08:05.328484 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.329825 4718 scope.go:117] "RemoveContainer" containerID="12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.343844 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.343883 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.343898 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.343918 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.343933 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.452876 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.452939 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.452959 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.452984 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.453012 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.555808 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.555866 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.555881 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.555903 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.555918 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.659611 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.659675 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.659693 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.659719 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.659737 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.762512 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.762579 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.762597 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.762622 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.762640 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.865652 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.865706 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.865722 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.865748 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.865766 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.969517 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.969562 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.969575 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.969592 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:05 crc kubenswrapper[4718]: I1206 04:08:05.969601 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:05Z","lastTransitionTime":"2025-12-06T04:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.072983 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.073042 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.073079 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.073100 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.073113 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:06Z","lastTransitionTime":"2025-12-06T04:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.175973 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.176009 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.176018 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.176037 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.176049 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:06Z","lastTransitionTime":"2025-12-06T04:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.278590 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.278631 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.278640 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.278660 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.278672 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:06Z","lastTransitionTime":"2025-12-06T04:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.328175 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:06 crc kubenswrapper[4718]: E1206 04:08:06.328419 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.381128 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.381182 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.381192 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.381210 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.381221 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:06Z","lastTransitionTime":"2025-12-06T04:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.483935 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.483975 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.483985 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.484003 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.484013 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:06Z","lastTransitionTime":"2025-12-06T04:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.587137 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.587487 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.587732 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.587941 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.588152 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:06Z","lastTransitionTime":"2025-12-06T04:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.691850 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.691913 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.691927 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.691953 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.691968 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:06Z","lastTransitionTime":"2025-12-06T04:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.767407 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/2.log" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.771148 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.771644 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.793085 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.796144 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.796273 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.796302 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.796338 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.796365 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:06Z","lastTransitionTime":"2025-12-06T04:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.808731 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.833188 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.855977 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.878188 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.900018 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.900098 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.900122 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.900174 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.900196 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:06Z","lastTransitionTime":"2025-12-06T04:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.904021 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.924474 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.941332 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.956201 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.974818 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:06 crc kubenswrapper[4718]: I1206 04:08:06.996393 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:59Z\\\",\\\"message\\\":\\\"2025-12-06T04:07:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389\\\\n2025-12-06T04:07:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389 to /host/opt/cni/bin/\\\\n2025-12-06T04:07:14Z [verbose] multus-daemon started\\\\n2025-12-06T04:07:14Z [verbose] Readiness Indicator file check\\\\n2025-12-06T04:07:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:06Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.003667 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.003746 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.003776 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.003808 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.003832 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.015832 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.035984 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.053496 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.071576 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.090154 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cc06da8-9a76-4ff2-be80-948fd12cc089\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dee1caa09708aba1318dc9c5f58cae75a93d70cbeb302d6b72b62d3842a1066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.111400 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.111466 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.111494 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.111520 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.111539 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.113312 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.126764 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.216303 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.216359 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.216375 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.216395 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.216414 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.322033 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.322083 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.322283 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.322318 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.322333 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.328216 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.329223 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.329288 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:07 crc kubenswrapper[4718]: E1206 04:08:07.329440 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:07 crc kubenswrapper[4718]: E1206 04:08:07.329915 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:07 crc kubenswrapper[4718]: E1206 04:08:07.330365 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.426688 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.426746 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.426763 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.426787 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.426806 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.530210 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.530339 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.530358 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.530381 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.530403 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.632773 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.632812 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.632824 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.632844 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.632855 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.736123 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.736199 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.736223 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.736309 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.736338 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.777689 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/3.log" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.778608 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/2.log" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.782488 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" exitCode=1 Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.782577 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.782696 4718 scope.go:117] "RemoveContainer" containerID="12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.783589 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:08:07 crc kubenswrapper[4718]: E1206 04:08:07.783874 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.806726 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.828052 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.839140 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.839263 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.839282 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.839310 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.839328 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.850529 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.865347 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.888667 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12ab28664ac3801897a06da2edeb15762cd6e9a6201202c52985e01525610463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:39Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 04:07:39.297983 6347 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 04:07:39.298583 6347 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 04:07:39.298621 6347 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 04:07:39.298647 6347 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 04:07:39.298666 6347 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 04:07:39.298671 6347 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 04:07:39.298892 6347 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 04:07:39.298893 6347 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 04:07:39.298921 6347 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 04:07:39.298931 6347 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 04:07:39.298956 6347 factory.go:656] Stopping watch factory\\\\nI1206 04:07:39.298972 6347 ovnkube.go:599] Stopped ovnkube\\\\nI1206 04:07:39.298989 6347 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 04:07:39.298999 6347 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:08:07Z\\\",\\\"message\\\":\\\"ce-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1206 04:08:06.547497 6696 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webh\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.905935 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.924310 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:59Z\\\",\\\"message\\\":\\\"2025-12-06T04:07:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389\\\\n2025-12-06T04:07:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389 to /host/opt/cni/bin/\\\\n2025-12-06T04:07:14Z [verbose] multus-daemon started\\\\n2025-12-06T04:07:14Z [verbose] Readiness Indicator file check\\\\n2025-12-06T04:07:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.938715 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.941830 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.941868 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.941879 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.941908 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.941921 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:07Z","lastTransitionTime":"2025-12-06T04:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.954989 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.977876 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:07 crc kubenswrapper[4718]: I1206 04:08:07.990263 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.001221 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:07Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.012025 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.028375 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.044709 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.044751 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.044764 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.044782 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.044795 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.046918 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cc06da8-9a76-4ff2-be80-948fd12cc089\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dee1caa09708aba1318dc9c5f58cae75a93d70cbeb302d6b72b62d3842a1066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.069272 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.086186 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.100607 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.148016 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.148200 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.148305 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.148374 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.148435 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.251744 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.251794 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.251806 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.251824 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.251838 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.327852 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:08 crc kubenswrapper[4718]: E1206 04:08:08.328035 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.355333 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.355374 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.355386 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.355407 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.355420 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.458094 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.458153 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.458170 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.458196 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.458215 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.560562 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.560657 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.560677 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.560700 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.560717 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.664019 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.664083 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.664100 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.664126 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.664143 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.767659 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.767712 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.767724 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.767743 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.767756 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.788083 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/3.log" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.792622 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:08:08 crc kubenswrapper[4718]: E1206 04:08:08.792893 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.810366 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.825369 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.845361 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.869014 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.871512 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.871576 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.871589 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.871606 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.871616 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.885904 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.900631 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.916347 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.932674 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:59Z\\\",\\\"message\\\":\\\"2025-12-06T04:07:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389\\\\n2025-12-06T04:07:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389 to /host/opt/cni/bin/\\\\n2025-12-06T04:07:14Z [verbose] multus-daemon started\\\\n2025-12-06T04:07:14Z [verbose] Readiness Indicator file check\\\\n2025-12-06T04:07:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.948016 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.961187 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cc06da8-9a76-4ff2-be80-948fd12cc089\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dee1caa09708aba1318dc9c5f58cae75a93d70cbeb302d6b72b62d3842a1066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.974477 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.974562 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.974584 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.974614 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.974639 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:08Z","lastTransitionTime":"2025-12-06T04:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:08 crc kubenswrapper[4718]: I1206 04:08:08.985071 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:08Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.003952 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:09Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.018837 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:09Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.034585 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:09Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.050105 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:09Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.069452 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:09Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.077096 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.077147 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.077156 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.077176 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.077187 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:09Z","lastTransitionTime":"2025-12-06T04:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.081749 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:09Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.107275 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:08:07Z\\\",\\\"message\\\":\\\"ce-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1206 04:08:06.547497 6696 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webh\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:08:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:09Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.181518 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.181591 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.181608 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.181636 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.181654 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:09Z","lastTransitionTime":"2025-12-06T04:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.284733 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.284806 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.284824 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.284902 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.284924 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:09Z","lastTransitionTime":"2025-12-06T04:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.327868 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.327901 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:09 crc kubenswrapper[4718]: E1206 04:08:09.328020 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.328088 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:09 crc kubenswrapper[4718]: E1206 04:08:09.328163 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:09 crc kubenswrapper[4718]: E1206 04:08:09.328352 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.387849 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.387962 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.388025 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.388057 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.388076 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:09Z","lastTransitionTime":"2025-12-06T04:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.491461 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.491508 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.491520 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.491537 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.491550 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:09Z","lastTransitionTime":"2025-12-06T04:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.594277 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.594327 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.594341 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.594358 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.594370 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:09Z","lastTransitionTime":"2025-12-06T04:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.698101 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.698156 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.698173 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.698196 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.698213 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:09Z","lastTransitionTime":"2025-12-06T04:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.801171 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.801259 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.801285 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.801315 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.801339 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:09Z","lastTransitionTime":"2025-12-06T04:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.905071 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.905138 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.905153 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.905177 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:09 crc kubenswrapper[4718]: I1206 04:08:09.905189 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:09Z","lastTransitionTime":"2025-12-06T04:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.008723 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.008794 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.008816 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.008849 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.008870 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.111513 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.111827 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.111972 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.112108 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.112227 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.214830 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.215322 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.215553 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.215762 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.215950 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.319376 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.319443 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.319455 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.319480 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.319496 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.327855 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:10 crc kubenswrapper[4718]: E1206 04:08:10.328077 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.422076 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.422145 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.422160 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.422178 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.422190 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.524770 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.524824 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.524839 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.524859 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.524873 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.627859 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.627917 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.627930 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.627963 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.627978 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.730996 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.731031 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.731042 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.731059 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.731071 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.833850 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.833906 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.833924 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.833951 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.833970 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.936829 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.936893 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.936912 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.936937 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:10 crc kubenswrapper[4718]: I1206 04:08:10.936956 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:10Z","lastTransitionTime":"2025-12-06T04:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.040133 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.040197 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.040215 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.040269 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.040290 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.142715 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.143051 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.143219 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.143422 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.143561 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.246497 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.246576 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.246595 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.246619 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.246636 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.327866 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.327916 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.327930 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:11 crc kubenswrapper[4718]: E1206 04:08:11.328218 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:11 crc kubenswrapper[4718]: E1206 04:08:11.328522 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:11 crc kubenswrapper[4718]: E1206 04:08:11.328375 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.343593 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cc06da8-9a76-4ff2-be80-948fd12cc089\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dee1caa09708aba1318dc9c5f58cae75a93d70cbeb302d6b72b62d3842a1066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e02d08b8cceaab9cbd9cc778ba18e0e70a3eeee33e6eb099845a8d0e9b13828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.350802 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.350919 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.350944 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.350974 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.350999 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.367589 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606ea280-e52e-4894-86cd-454e3f550227\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f3ddca49c78a1b952ba6775a0a0d1876291e21110a53f2d07f3b33654d4ca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c9aca6007d4f225d3a82afc85f40a8cfd130d918037226b6664649ec0d7ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a69a7d49260e308e85261da7d7e29a69ff93d09993a3d2f67a49244e8ac4343\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://298541740ef68b67fc20ff310583b1e6a0a24f8e3241fd2b475c43510302591e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88dabfbe7375a8cb5bc35de01463818b4ed24f8f77ede4fe203d327a1070cc95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0162a213b3b0769852400039b4aaa4e91ba618356f49730d49e618ff3b777c13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3918f817101f100c07c37da1aa1de3530da9837966092da28342e4f41129065\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2bmp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jf6pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.386884 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a81cbde5-9395-418f-a205-279a33268d9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d8cf57bd148d181101397fb7a026652ba03dc1f5468317403883cc53b8260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44133d4298bc8af05d33686c4fd572d6520d2a1c026850ef4cf6eb46eb6529c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xch99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7d4cs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.402175 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fxdkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10ed0065-fa7b-4190-b140-ddcf6c368ef9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656cca1a1fa506531b05ead00da01b17f9475f0cf97a21f27b0bdcc74e4058b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljm9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fxdkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.420582 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67f564b0734e1f5a68d6eaf157012cd8372ce837109ebf6acbd99cd1169167d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.439739 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.453407 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.453473 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.453493 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.453522 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.453545 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.456254 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.467884 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9gdpg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b4d7e05-3191-4d42-adc8-b07d91e52191\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6bce52b41dc215ae664e25101929fd9b6047b7972168ea10d56b8ed67b936aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z52pb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9gdpg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.489002 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9ed5db1-456b-4891-a750-809728e9c0a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:08:07Z\\\",\\\"message\\\":\\\"ce-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1206 04:08:06.547497 6696 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webh\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:08:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qnpgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6nfdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.503125 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87a3c869-d0a2-46cd-ac46-10022d92c7af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6ce2d3c458e149ffd3f37f983b4f88f9fe1a3656427854795ceda13712449ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpbx5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pr7fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.516935 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4r7xc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2272fbce-ce87-4f9f-b1c5-f83f262534de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T04:07:59Z\\\",\\\"message\\\":\\\"2025-12-06T04:07:14+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389\\\\n2025-12-06T04:07:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b404b306-f756-4198-bcfb-bc0cb9f90389 to /host/opt/cni/bin/\\\\n2025-12-06T04:07:14Z [verbose] multus-daemon started\\\\n2025-12-06T04:07:14Z [verbose] Readiness Indicator file check\\\\n2025-12-06T04:07:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:07:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hz2nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4r7xc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.530606 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee82ee8a-a2d6-4dca-a959-87c5e2b55b4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3a287266d93fc98a87cf442ac6ff187a2b799245570ead6098afb02a905205d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec737c1edf40538875d74b641c2b66c9214c88d30dd882d7dbcb510945a23e6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a1355525412e64f327619fc166bef7b624d08159d6133837623957f5993c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299f2cb21b60c3627fcc61266db4f5cf58ad59b197a88510b03115f8f19abfe7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.546500 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3051b71d-5496-4229-8a3e-b7875ed28a01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4b7351df0620da5dc40986eb0ba5876ea123bf7e0a8f5589551c2a945336d73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5656ac784c3fef99c0b799f97cd97d03500fae5973994577dda881f1a6dd153\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b62a564a38a9bd231697d9a0ff4eee294e952e2256454ba858659d8107247e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.556324 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.556379 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.556392 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.556418 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.556434 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.563632 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"245fe818-e203-4799-93f1-1f96b6972225\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 04:07:03.756121 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 04:07:03.757040 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2922789406/tls.crt::/tmp/serving-cert-2922789406/tls.key\\\\\\\"\\\\nI1206 04:07:09.281065 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 04:07:09.288881 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 04:07:09.288959 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 04:07:09.289006 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 04:07:09.289031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 04:07:09.312646 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 04:07:09.312697 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312712 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 04:07:09.312725 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 04:07:09.312734 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 04:07:09.312742 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 04:07:09.312749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 04:07:09.313149 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 04:07:09.317849 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T04:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T04:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.579011 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.593608 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbbb024e0a556b72b0047488d8a581993570d64e82248f6581f071ae03e2e59a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd7c9377881ff96f6f4357f1d114c08d473ac2760181876edb9e00956282a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.610252 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959c44e68558eb876d5f4775e5bb865b75a67ef40983a9abb85067abb2ec9b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T04:07:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.627369 4718 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38892dbb-af9b-4483-bfe6-c1ed160d356c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T04:07:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jlm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T04:07:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mhz2w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:11Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.659477 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.659512 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.659522 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.659535 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.659544 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.761834 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.761890 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.761907 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.761930 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.761946 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.865538 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.865591 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.865615 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.865638 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.865655 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.968984 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.969428 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.969447 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.969470 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:11 crc kubenswrapper[4718]: I1206 04:08:11.969490 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:11Z","lastTransitionTime":"2025-12-06T04:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.073383 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.073451 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.073471 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.073496 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.073517 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:12Z","lastTransitionTime":"2025-12-06T04:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.177431 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.177515 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.177538 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.177571 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.177592 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:12Z","lastTransitionTime":"2025-12-06T04:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.280326 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.280754 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.280946 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.281144 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.281379 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:12Z","lastTransitionTime":"2025-12-06T04:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.328062 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:12 crc kubenswrapper[4718]: E1206 04:08:12.328212 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.384203 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.384266 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.384280 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.384296 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.384308 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:12Z","lastTransitionTime":"2025-12-06T04:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.487929 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.487982 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.487993 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.488014 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.488028 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:12Z","lastTransitionTime":"2025-12-06T04:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.592160 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.592277 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.592299 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.592327 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.592347 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:12Z","lastTransitionTime":"2025-12-06T04:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.695997 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.696075 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.696095 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.696121 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.696140 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:12Z","lastTransitionTime":"2025-12-06T04:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.799705 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.799766 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.799783 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.799806 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.799826 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:12Z","lastTransitionTime":"2025-12-06T04:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.904118 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.904185 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.904203 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.904227 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:12 crc kubenswrapper[4718]: I1206 04:08:12.904272 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:12Z","lastTransitionTime":"2025-12-06T04:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.007605 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.007677 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.007709 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.007742 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.007766 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.110896 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.110965 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.110982 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.111476 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.111530 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.139563 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.139738 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:17.139704474 +0000 UTC m=+146.145409675 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.139805 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.139861 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.139932 4718 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.140002 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:09:17.139984246 +0000 UTC m=+146.145689447 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.140033 4718 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.140086 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 04:09:17.14007229 +0000 UTC m=+146.145777481 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.214518 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.214597 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.214621 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.214651 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.214674 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.241556 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.241707 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.241845 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.242077 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.242102 4718 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.241918 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.242207 4718 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.242242 4718 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.242173 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 04:09:17.242151566 +0000 UTC m=+146.247856767 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.242364 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 04:09:17.242337274 +0000 UTC m=+146.248042435 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.317583 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.317671 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.317697 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.317729 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.317748 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.327190 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.327369 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.327438 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.327473 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.327662 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.327766 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.421611 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.421676 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.421696 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.421720 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.421808 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.524822 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.524871 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.524882 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.524898 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.524909 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.627371 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.627451 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.627474 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.627503 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.627522 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.731485 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.731555 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.731573 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.731599 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.731616 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.787023 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.787070 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.787086 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.787112 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.787129 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.808991 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.813909 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.813960 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.813977 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.814001 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.814018 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.834573 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.839182 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.839216 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.839224 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.839262 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.839272 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.857851 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.862114 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.862174 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.862195 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.862220 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.862293 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.881580 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.887060 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.887136 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.887160 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.887184 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.887201 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.908572 4718 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T04:08:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3f36729d-94a7-4ff7-b13e-5910bb9e8ea4\\\",\\\"systemUUID\\\":\\\"76af1561-f0fa-462c-8fb7-968628565864\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T04:08:13Z is after 2025-08-24T17:21:41Z" Dec 06 04:08:13 crc kubenswrapper[4718]: E1206 04:08:13.908726 4718 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.910876 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.910897 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.910908 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.910924 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:13 crc kubenswrapper[4718]: I1206 04:08:13.910936 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:13Z","lastTransitionTime":"2025-12-06T04:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.013568 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.013614 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.013630 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.013653 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.013671 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.117216 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.117286 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.117302 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.117325 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.117341 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.220388 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.220726 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.220866 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.221047 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.221195 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.324316 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.324367 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.324382 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.324402 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.324417 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.327948 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:14 crc kubenswrapper[4718]: E1206 04:08:14.328184 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.427915 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.427957 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.427968 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.427986 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.428000 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.531721 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.531768 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.531784 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.531806 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.531823 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.634991 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.635072 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.635092 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.635120 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.635137 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.737979 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.738083 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.738101 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.738127 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.738184 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.841355 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.841409 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.841425 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.841448 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.841465 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.944603 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.944950 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.945106 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.945349 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:14 crc kubenswrapper[4718]: I1206 04:08:14.945497 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:14Z","lastTransitionTime":"2025-12-06T04:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.048475 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.049457 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.049525 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.049556 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.049576 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.152470 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.152512 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.152525 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.152541 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.152552 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.256139 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.256208 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.256228 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.256302 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.256324 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.328526 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.328561 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.328633 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:15 crc kubenswrapper[4718]: E1206 04:08:15.328845 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:15 crc kubenswrapper[4718]: E1206 04:08:15.328956 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:15 crc kubenswrapper[4718]: E1206 04:08:15.329113 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.359130 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.359190 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.359218 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.359299 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.359323 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.462542 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.462598 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.462615 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.462638 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.462655 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.564911 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.564960 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.564976 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.564996 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.565011 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.668795 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.668954 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.668981 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.669012 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.669034 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.772499 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.772555 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.772572 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.772595 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.772613 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.875449 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.875525 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.875547 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.875583 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.875606 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.979054 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.979465 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.979494 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.979524 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:15 crc kubenswrapper[4718]: I1206 04:08:15.979548 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:15Z","lastTransitionTime":"2025-12-06T04:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.083199 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.083300 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.083320 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.083346 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.083364 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:16Z","lastTransitionTime":"2025-12-06T04:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.186974 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.187055 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.187079 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.187104 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.187123 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:16Z","lastTransitionTime":"2025-12-06T04:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.290480 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.290541 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.290560 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.290585 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.290600 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:16Z","lastTransitionTime":"2025-12-06T04:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.327757 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:16 crc kubenswrapper[4718]: E1206 04:08:16.328093 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.393623 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.393934 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.394004 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.394023 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.394035 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:16Z","lastTransitionTime":"2025-12-06T04:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.498296 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.498374 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.498399 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.498433 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.498455 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:16Z","lastTransitionTime":"2025-12-06T04:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.601878 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.602362 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.602599 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.602815 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.602979 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:16Z","lastTransitionTime":"2025-12-06T04:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.706028 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.706108 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.706134 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.706165 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.706183 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:16Z","lastTransitionTime":"2025-12-06T04:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.810096 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.810175 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.810196 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.810226 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.810282 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:16Z","lastTransitionTime":"2025-12-06T04:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.913324 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.913387 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.913403 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.913427 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:16 crc kubenswrapper[4718]: I1206 04:08:16.913446 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:16Z","lastTransitionTime":"2025-12-06T04:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.016811 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.016875 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.016892 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.016932 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.016953 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.125863 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.126005 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.126025 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.126049 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.126103 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.229632 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.229778 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.229798 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.229819 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.229833 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.328196 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.328213 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:17 crc kubenswrapper[4718]: E1206 04:08:17.328463 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.328421 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:17 crc kubenswrapper[4718]: E1206 04:08:17.328604 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:17 crc kubenswrapper[4718]: E1206 04:08:17.328734 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.332109 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.332165 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.332188 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.332219 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.332775 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.437066 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.437128 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.437146 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.437169 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.437186 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.540268 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.540318 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.540336 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.540360 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.540377 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.644270 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.644573 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.644669 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.644772 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.644866 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.748391 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.748481 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.748501 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.748523 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.748572 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.851324 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.851385 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.851403 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.851453 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.851472 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.954890 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.954956 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.954974 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.954999 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:17 crc kubenswrapper[4718]: I1206 04:08:17.955017 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:17Z","lastTransitionTime":"2025-12-06T04:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.057868 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.057895 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.057906 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.057919 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.057931 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.160772 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.160837 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.160860 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.160891 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.160910 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.264547 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.264616 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.264629 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.264678 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.264692 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.327891 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:18 crc kubenswrapper[4718]: E1206 04:08:18.328144 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.368719 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.368762 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.368774 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.368792 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.368804 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.471413 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.471446 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.471455 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.471468 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.471476 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.574159 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.574429 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.574454 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.574484 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.574504 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.677764 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.677829 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.677853 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.677918 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.677938 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.781046 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.781106 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.781124 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.781151 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.781171 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.884820 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.884857 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.884868 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.884883 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.884898 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.987414 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.987470 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.987489 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.987517 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:18 crc kubenswrapper[4718]: I1206 04:08:18.987537 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:18Z","lastTransitionTime":"2025-12-06T04:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.090269 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.090313 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.090324 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.090340 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.090351 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:19Z","lastTransitionTime":"2025-12-06T04:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.193875 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.193936 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.193954 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.193981 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.194001 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:19Z","lastTransitionTime":"2025-12-06T04:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.296551 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.296611 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.296627 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.296649 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.296665 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:19Z","lastTransitionTime":"2025-12-06T04:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.328123 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.328160 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.328219 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:19 crc kubenswrapper[4718]: E1206 04:08:19.328339 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:19 crc kubenswrapper[4718]: E1206 04:08:19.328462 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:19 crc kubenswrapper[4718]: E1206 04:08:19.328606 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.398634 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.398670 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.398682 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.398696 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.398709 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:19Z","lastTransitionTime":"2025-12-06T04:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.501764 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.501818 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.501837 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.501867 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.501893 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:19Z","lastTransitionTime":"2025-12-06T04:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.604884 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.604947 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.604964 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.604986 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.605009 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:19Z","lastTransitionTime":"2025-12-06T04:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.706977 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.707028 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.707046 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.707069 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.707085 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:19Z","lastTransitionTime":"2025-12-06T04:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.810486 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.810598 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.810617 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.810676 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.810693 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:19Z","lastTransitionTime":"2025-12-06T04:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.920662 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.920739 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.920759 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.920785 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:19 crc kubenswrapper[4718]: I1206 04:08:19.920806 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:19Z","lastTransitionTime":"2025-12-06T04:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.023662 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.023714 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.023726 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.023744 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.023757 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.126938 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.126977 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.126988 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.127002 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.127014 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.230561 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.230622 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.230638 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.230665 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.230687 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.327815 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:20 crc kubenswrapper[4718]: E1206 04:08:20.327974 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.333795 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.333840 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.333854 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.333870 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.333879 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.436595 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.436640 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.436651 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.436668 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.436679 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.539879 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.539921 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.539933 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.539952 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.539964 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.643014 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.643074 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.643091 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.643120 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.643137 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.746759 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.746826 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.746844 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.746870 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.746888 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.850716 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.850793 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.850815 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.850839 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.850858 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.953163 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.953222 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.953273 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.953299 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:20 crc kubenswrapper[4718]: I1206 04:08:20.953317 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:20Z","lastTransitionTime":"2025-12-06T04:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.056819 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.056883 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.056901 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.056925 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.056941 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.162160 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.162269 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.162297 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.162326 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.162346 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.264864 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.264903 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.264913 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.264928 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.264938 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.328110 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.328190 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.328284 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:21 crc kubenswrapper[4718]: E1206 04:08:21.328408 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:21 crc kubenswrapper[4718]: E1206 04:08:21.328730 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:21 crc kubenswrapper[4718]: E1206 04:08:21.328885 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.351723 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=40.351700269 podStartE2EDuration="40.351700269s" podCreationTimestamp="2025-12-06 04:07:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.349587723 +0000 UTC m=+90.355292924" watchObservedRunningTime="2025-12-06 04:08:21.351700269 +0000 UTC m=+90.357405440" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.368387 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.368369916 podStartE2EDuration="1m8.368369916s" podCreationTimestamp="2025-12-06 04:07:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.368080643 +0000 UTC m=+90.373785824" watchObservedRunningTime="2025-12-06 04:08:21.368369916 +0000 UTC m=+90.374075077" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.373512 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.373538 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.373548 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.373559 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.373569 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.396696 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=72.39663859 podStartE2EDuration="1m12.39663859s" podCreationTimestamp="2025-12-06 04:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.395397713 +0000 UTC m=+90.401102894" watchObservedRunningTime="2025-12-06 04:08:21.39663859 +0000 UTC m=+90.402343761" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.466395 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podStartSLOduration=71.466366927 podStartE2EDuration="1m11.466366927s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.466259612 +0000 UTC m=+90.471964783" watchObservedRunningTime="2025-12-06 04:08:21.466366927 +0000 UTC m=+90.472072128" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.476155 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.476214 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.476255 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.476281 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.476299 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.488140 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-4r7xc" podStartSLOduration=71.488119854 podStartE2EDuration="1m11.488119854s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.487949376 +0000 UTC m=+90.493654537" watchObservedRunningTime="2025-12-06 04:08:21.488119854 +0000 UTC m=+90.493825015" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.513098 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=20.513079708 podStartE2EDuration="20.513079708s" podCreationTimestamp="2025-12-06 04:08:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.512553174 +0000 UTC m=+90.518258335" watchObservedRunningTime="2025-12-06 04:08:21.513079708 +0000 UTC m=+90.518784879" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.531257 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-jf6pj" podStartSLOduration=71.531242113 podStartE2EDuration="1m11.531242113s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.531168899 +0000 UTC m=+90.536874100" watchObservedRunningTime="2025-12-06 04:08:21.531242113 +0000 UTC m=+90.536947274" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.545000 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7d4cs" podStartSLOduration=71.544969206 podStartE2EDuration="1m11.544969206s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.543980811 +0000 UTC m=+90.549685972" watchObservedRunningTime="2025-12-06 04:08:21.544969206 +0000 UTC m=+90.550674407" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.555312 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-fxdkw" podStartSLOduration=71.555284944 podStartE2EDuration="1m11.555284944s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.55496319 +0000 UTC m=+90.560668421" watchObservedRunningTime="2025-12-06 04:08:21.555284944 +0000 UTC m=+90.560990145" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.579691 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.579753 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.579771 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.579797 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.579818 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.655895 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-9gdpg" podStartSLOduration=71.655870712 podStartE2EDuration="1m11.655870712s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:21.625688482 +0000 UTC m=+90.631393673" watchObservedRunningTime="2025-12-06 04:08:21.655870712 +0000 UTC m=+90.661575903" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.682911 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.682960 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.682969 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.682985 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.682994 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.785791 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.785855 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.785872 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.785894 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.785913 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.888898 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.888948 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.888967 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.888991 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.889009 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.992422 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.992496 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.992515 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.992540 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:21 crc kubenswrapper[4718]: I1206 04:08:21.992558 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:21Z","lastTransitionTime":"2025-12-06T04:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.096036 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.096085 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.096096 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.096114 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.096125 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:22Z","lastTransitionTime":"2025-12-06T04:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.198773 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.199070 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.199158 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.199303 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.199417 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:22Z","lastTransitionTime":"2025-12-06T04:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.303020 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.303097 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.303122 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.303152 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.303173 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:22Z","lastTransitionTime":"2025-12-06T04:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.327211 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:22 crc kubenswrapper[4718]: E1206 04:08:22.327474 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.348590 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.406280 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.406342 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.406361 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.406385 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.406403 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:22Z","lastTransitionTime":"2025-12-06T04:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.509393 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.509687 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.509780 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.509862 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.509956 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:22Z","lastTransitionTime":"2025-12-06T04:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.613600 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.613671 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.613692 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.613723 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.613743 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:22Z","lastTransitionTime":"2025-12-06T04:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.717696 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.717743 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.717781 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.717797 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.717809 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:22Z","lastTransitionTime":"2025-12-06T04:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.820377 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.820436 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.820456 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.820483 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.820505 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:22Z","lastTransitionTime":"2025-12-06T04:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.923393 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.923453 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.923471 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.923527 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:22 crc kubenswrapper[4718]: I1206 04:08:22.923545 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:22Z","lastTransitionTime":"2025-12-06T04:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.026524 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.026578 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.026593 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.026616 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.026635 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.129954 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.130015 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.130036 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.130063 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.130084 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.232111 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.232575 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.232747 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.232894 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.233038 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.327368 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.327368 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:23 crc kubenswrapper[4718]: E1206 04:08:23.327514 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:23 crc kubenswrapper[4718]: E1206 04:08:23.327634 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.327368 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:23 crc kubenswrapper[4718]: E1206 04:08:23.327749 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.335097 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.335147 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.335166 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.335191 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.335209 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.438765 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.438856 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.438928 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.438961 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.439039 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.541940 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.541993 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.542015 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.542043 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.542065 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.645194 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.645307 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.645330 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.645361 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.645382 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.749372 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.749476 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.749504 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.749574 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.749591 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.851066 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.851133 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.851147 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.851164 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.851175 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.953585 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.953630 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.953642 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.953663 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:23 crc kubenswrapper[4718]: I1206 04:08:23.953680 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:23Z","lastTransitionTime":"2025-12-06T04:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.056117 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.056562 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.056641 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.056718 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.056789 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:24Z","lastTransitionTime":"2025-12-06T04:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.121687 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.121918 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.122021 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.122108 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.122182 4718 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T04:08:24Z","lastTransitionTime":"2025-12-06T04:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.164909 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k"] Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.165316 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.168798 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.169292 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.170540 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.171891 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.200419 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=2.200378277 podStartE2EDuration="2.200378277s" podCreationTimestamp="2025-12-06 04:08:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:24.200194159 +0000 UTC m=+93.205899320" watchObservedRunningTime="2025-12-06 04:08:24.200378277 +0000 UTC m=+93.206083448" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.272510 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/86ae9954-8dbb-4670-964b-3d8406839517-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.272740 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/86ae9954-8dbb-4670-964b-3d8406839517-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.272827 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/86ae9954-8dbb-4670-964b-3d8406839517-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.272978 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/86ae9954-8dbb-4670-964b-3d8406839517-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.273070 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86ae9954-8dbb-4670-964b-3d8406839517-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.327887 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:24 crc kubenswrapper[4718]: E1206 04:08:24.328069 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.328893 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:08:24 crc kubenswrapper[4718]: E1206 04:08:24.329093 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.374485 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/86ae9954-8dbb-4670-964b-3d8406839517-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.374521 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/86ae9954-8dbb-4670-964b-3d8406839517-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.374544 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/86ae9954-8dbb-4670-964b-3d8406839517-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.374626 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/86ae9954-8dbb-4670-964b-3d8406839517-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.374642 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/86ae9954-8dbb-4670-964b-3d8406839517-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.374661 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86ae9954-8dbb-4670-964b-3d8406839517-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.374936 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/86ae9954-8dbb-4670-964b-3d8406839517-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.380138 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/86ae9954-8dbb-4670-964b-3d8406839517-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.382254 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86ae9954-8dbb-4670-964b-3d8406839517-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.392702 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/86ae9954-8dbb-4670-964b-3d8406839517-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xm44k\" (UID: \"86ae9954-8dbb-4670-964b-3d8406839517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.480902 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" Dec 06 04:08:24 crc kubenswrapper[4718]: I1206 04:08:24.853567 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" event={"ID":"86ae9954-8dbb-4670-964b-3d8406839517","Type":"ContainerStarted","Data":"b1d6d96b838824a1f2f94112549307b851f555575eac1d1d8841a1d587ec57a2"} Dec 06 04:08:25 crc kubenswrapper[4718]: I1206 04:08:25.328115 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:25 crc kubenswrapper[4718]: I1206 04:08:25.328160 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:25 crc kubenswrapper[4718]: I1206 04:08:25.328206 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:25 crc kubenswrapper[4718]: E1206 04:08:25.328345 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:25 crc kubenswrapper[4718]: E1206 04:08:25.328574 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:25 crc kubenswrapper[4718]: E1206 04:08:25.328889 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:26 crc kubenswrapper[4718]: I1206 04:08:26.328192 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:26 crc kubenswrapper[4718]: E1206 04:08:26.328409 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:27 crc kubenswrapper[4718]: I1206 04:08:27.327973 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:27 crc kubenswrapper[4718]: I1206 04:08:27.328055 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:27 crc kubenswrapper[4718]: E1206 04:08:27.328152 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:27 crc kubenswrapper[4718]: E1206 04:08:27.328356 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:27 crc kubenswrapper[4718]: I1206 04:08:27.328484 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:27 crc kubenswrapper[4718]: E1206 04:08:27.328573 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:28 crc kubenswrapper[4718]: I1206 04:08:28.328393 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:28 crc kubenswrapper[4718]: E1206 04:08:28.328899 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:28 crc kubenswrapper[4718]: I1206 04:08:28.875461 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" event={"ID":"86ae9954-8dbb-4670-964b-3d8406839517","Type":"ContainerStarted","Data":"73d745026e3b587fd2280a0017d796490cf758433d2bf14f50dfef0865efcb13"} Dec 06 04:08:28 crc kubenswrapper[4718]: I1206 04:08:28.892546 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xm44k" podStartSLOduration=78.892514874 podStartE2EDuration="1m18.892514874s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:28.891662205 +0000 UTC m=+97.897367356" watchObservedRunningTime="2025-12-06 04:08:28.892514874 +0000 UTC m=+97.898220075" Dec 06 04:08:29 crc kubenswrapper[4718]: I1206 04:08:29.133178 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:29 crc kubenswrapper[4718]: E1206 04:08:29.133499 4718 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:08:29 crc kubenswrapper[4718]: E1206 04:08:29.133673 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs podName:38892dbb-af9b-4483-bfe6-c1ed160d356c nodeName:}" failed. No retries permitted until 2025-12-06 04:09:33.133604683 +0000 UTC m=+162.139309884 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs") pod "network-metrics-daemon-mhz2w" (UID: "38892dbb-af9b-4483-bfe6-c1ed160d356c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 04:08:29 crc kubenswrapper[4718]: I1206 04:08:29.327675 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:29 crc kubenswrapper[4718]: I1206 04:08:29.327700 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:29 crc kubenswrapper[4718]: E1206 04:08:29.327871 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:29 crc kubenswrapper[4718]: I1206 04:08:29.327702 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:29 crc kubenswrapper[4718]: E1206 04:08:29.327948 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:29 crc kubenswrapper[4718]: E1206 04:08:29.328069 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:30 crc kubenswrapper[4718]: I1206 04:08:30.327473 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:30 crc kubenswrapper[4718]: E1206 04:08:30.327895 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:31 crc kubenswrapper[4718]: I1206 04:08:31.327450 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:31 crc kubenswrapper[4718]: I1206 04:08:31.327487 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:31 crc kubenswrapper[4718]: I1206 04:08:31.327537 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:31 crc kubenswrapper[4718]: E1206 04:08:31.328903 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:31 crc kubenswrapper[4718]: E1206 04:08:31.329062 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:31 crc kubenswrapper[4718]: E1206 04:08:31.329203 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:32 crc kubenswrapper[4718]: I1206 04:08:32.328048 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:32 crc kubenswrapper[4718]: E1206 04:08:32.328319 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:33 crc kubenswrapper[4718]: I1206 04:08:33.328094 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:33 crc kubenswrapper[4718]: I1206 04:08:33.328095 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:33 crc kubenswrapper[4718]: I1206 04:08:33.328137 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:33 crc kubenswrapper[4718]: E1206 04:08:33.328378 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:33 crc kubenswrapper[4718]: E1206 04:08:33.328500 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:33 crc kubenswrapper[4718]: E1206 04:08:33.328644 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:34 crc kubenswrapper[4718]: I1206 04:08:34.360950 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:34 crc kubenswrapper[4718]: E1206 04:08:34.361159 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:35 crc kubenswrapper[4718]: I1206 04:08:35.328353 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:35 crc kubenswrapper[4718]: I1206 04:08:35.328433 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:35 crc kubenswrapper[4718]: I1206 04:08:35.328380 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:35 crc kubenswrapper[4718]: E1206 04:08:35.328582 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:35 crc kubenswrapper[4718]: E1206 04:08:35.328699 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:35 crc kubenswrapper[4718]: E1206 04:08:35.328803 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:36 crc kubenswrapper[4718]: I1206 04:08:36.327309 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:36 crc kubenswrapper[4718]: E1206 04:08:36.327793 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:37 crc kubenswrapper[4718]: I1206 04:08:37.328519 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:37 crc kubenswrapper[4718]: I1206 04:08:37.328558 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:37 crc kubenswrapper[4718]: E1206 04:08:37.328773 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:37 crc kubenswrapper[4718]: E1206 04:08:37.328915 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:37 crc kubenswrapper[4718]: I1206 04:08:37.328571 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:37 crc kubenswrapper[4718]: E1206 04:08:37.329093 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:38 crc kubenswrapper[4718]: I1206 04:08:38.327884 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:38 crc kubenswrapper[4718]: E1206 04:08:38.328037 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:39 crc kubenswrapper[4718]: I1206 04:08:39.327552 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:39 crc kubenswrapper[4718]: I1206 04:08:39.327725 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:39 crc kubenswrapper[4718]: E1206 04:08:39.327874 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:39 crc kubenswrapper[4718]: I1206 04:08:39.328169 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:39 crc kubenswrapper[4718]: E1206 04:08:39.328447 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:39 crc kubenswrapper[4718]: E1206 04:08:39.329193 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:39 crc kubenswrapper[4718]: I1206 04:08:39.329649 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:08:39 crc kubenswrapper[4718]: E1206 04:08:39.329929 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6nfdl_openshift-ovn-kubernetes(e9ed5db1-456b-4891-a750-809728e9c0a5)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" Dec 06 04:08:40 crc kubenswrapper[4718]: I1206 04:08:40.327645 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:40 crc kubenswrapper[4718]: E1206 04:08:40.328057 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:41 crc kubenswrapper[4718]: I1206 04:08:41.328078 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:41 crc kubenswrapper[4718]: I1206 04:08:41.328104 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:41 crc kubenswrapper[4718]: E1206 04:08:41.330021 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:41 crc kubenswrapper[4718]: I1206 04:08:41.330069 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:41 crc kubenswrapper[4718]: E1206 04:08:41.330395 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:41 crc kubenswrapper[4718]: E1206 04:08:41.330501 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:42 crc kubenswrapper[4718]: I1206 04:08:42.327665 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:42 crc kubenswrapper[4718]: E1206 04:08:42.327848 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:43 crc kubenswrapper[4718]: I1206 04:08:43.327688 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:43 crc kubenswrapper[4718]: I1206 04:08:43.327780 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:43 crc kubenswrapper[4718]: E1206 04:08:43.327840 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:43 crc kubenswrapper[4718]: I1206 04:08:43.327714 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:43 crc kubenswrapper[4718]: E1206 04:08:43.328200 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:43 crc kubenswrapper[4718]: E1206 04:08:43.328395 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:44 crc kubenswrapper[4718]: I1206 04:08:44.327439 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:44 crc kubenswrapper[4718]: E1206 04:08:44.327651 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:45 crc kubenswrapper[4718]: I1206 04:08:45.327758 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:45 crc kubenswrapper[4718]: I1206 04:08:45.327821 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:45 crc kubenswrapper[4718]: I1206 04:08:45.327758 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:45 crc kubenswrapper[4718]: E1206 04:08:45.327989 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:45 crc kubenswrapper[4718]: E1206 04:08:45.328064 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:45 crc kubenswrapper[4718]: E1206 04:08:45.328212 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:45 crc kubenswrapper[4718]: I1206 04:08:45.943943 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4r7xc_2272fbce-ce87-4f9f-b1c5-f83f262534de/kube-multus/1.log" Dec 06 04:08:45 crc kubenswrapper[4718]: I1206 04:08:45.944674 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4r7xc_2272fbce-ce87-4f9f-b1c5-f83f262534de/kube-multus/0.log" Dec 06 04:08:45 crc kubenswrapper[4718]: I1206 04:08:45.944756 4718 generic.go:334] "Generic (PLEG): container finished" podID="2272fbce-ce87-4f9f-b1c5-f83f262534de" containerID="c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a" exitCode=1 Dec 06 04:08:45 crc kubenswrapper[4718]: I1206 04:08:45.944807 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4r7xc" event={"ID":"2272fbce-ce87-4f9f-b1c5-f83f262534de","Type":"ContainerDied","Data":"c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a"} Dec 06 04:08:45 crc kubenswrapper[4718]: I1206 04:08:45.944855 4718 scope.go:117] "RemoveContainer" containerID="7fab8c492f5b7528d855bb042fb7e3e7317a8dd2c83dce14d81ce9d38a5de621" Dec 06 04:08:45 crc kubenswrapper[4718]: I1206 04:08:45.946278 4718 scope.go:117] "RemoveContainer" containerID="c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a" Dec 06 04:08:45 crc kubenswrapper[4718]: E1206 04:08:45.946632 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-4r7xc_openshift-multus(2272fbce-ce87-4f9f-b1c5-f83f262534de)\"" pod="openshift-multus/multus-4r7xc" podUID="2272fbce-ce87-4f9f-b1c5-f83f262534de" Dec 06 04:08:46 crc kubenswrapper[4718]: I1206 04:08:46.327703 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:46 crc kubenswrapper[4718]: E1206 04:08:46.327847 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:46 crc kubenswrapper[4718]: I1206 04:08:46.952672 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4r7xc_2272fbce-ce87-4f9f-b1c5-f83f262534de/kube-multus/1.log" Dec 06 04:08:47 crc kubenswrapper[4718]: I1206 04:08:47.327815 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:47 crc kubenswrapper[4718]: E1206 04:08:47.328487 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:47 crc kubenswrapper[4718]: I1206 04:08:47.329029 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:47 crc kubenswrapper[4718]: E1206 04:08:47.329219 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:47 crc kubenswrapper[4718]: I1206 04:08:47.329668 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:47 crc kubenswrapper[4718]: E1206 04:08:47.329905 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:48 crc kubenswrapper[4718]: I1206 04:08:48.327778 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:48 crc kubenswrapper[4718]: E1206 04:08:48.327921 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:49 crc kubenswrapper[4718]: I1206 04:08:49.328200 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:49 crc kubenswrapper[4718]: I1206 04:08:49.328275 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:49 crc kubenswrapper[4718]: E1206 04:08:49.328437 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:49 crc kubenswrapper[4718]: I1206 04:08:49.328461 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:49 crc kubenswrapper[4718]: E1206 04:08:49.328577 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:49 crc kubenswrapper[4718]: E1206 04:08:49.328871 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:50 crc kubenswrapper[4718]: I1206 04:08:50.328354 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:50 crc kubenswrapper[4718]: E1206 04:08:50.328567 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:51 crc kubenswrapper[4718]: E1206 04:08:51.273355 4718 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 06 04:08:51 crc kubenswrapper[4718]: I1206 04:08:51.327528 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:51 crc kubenswrapper[4718]: E1206 04:08:51.329648 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:51 crc kubenswrapper[4718]: I1206 04:08:51.329689 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:51 crc kubenswrapper[4718]: I1206 04:08:51.329869 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:51 crc kubenswrapper[4718]: E1206 04:08:51.330079 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:51 crc kubenswrapper[4718]: E1206 04:08:51.330279 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:51 crc kubenswrapper[4718]: E1206 04:08:51.428729 4718 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 04:08:52 crc kubenswrapper[4718]: I1206 04:08:52.327583 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:52 crc kubenswrapper[4718]: E1206 04:08:52.328159 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:53 crc kubenswrapper[4718]: I1206 04:08:53.328092 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:53 crc kubenswrapper[4718]: I1206 04:08:53.328092 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:53 crc kubenswrapper[4718]: E1206 04:08:53.328353 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:53 crc kubenswrapper[4718]: I1206 04:08:53.328412 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:53 crc kubenswrapper[4718]: E1206 04:08:53.328551 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:53 crc kubenswrapper[4718]: E1206 04:08:53.328672 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:54 crc kubenswrapper[4718]: I1206 04:08:54.327795 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:54 crc kubenswrapper[4718]: E1206 04:08:54.328391 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:54 crc kubenswrapper[4718]: I1206 04:08:54.328976 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:08:54 crc kubenswrapper[4718]: I1206 04:08:54.985534 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/3.log" Dec 06 04:08:54 crc kubenswrapper[4718]: I1206 04:08:54.988443 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerStarted","Data":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} Dec 06 04:08:54 crc kubenswrapper[4718]: I1206 04:08:54.988930 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:08:55 crc kubenswrapper[4718]: I1206 04:08:55.013568 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podStartSLOduration=105.013548964 podStartE2EDuration="1m45.013548964s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:08:55.011885559 +0000 UTC m=+124.017590740" watchObservedRunningTime="2025-12-06 04:08:55.013548964 +0000 UTC m=+124.019254125" Dec 06 04:08:55 crc kubenswrapper[4718]: I1206 04:08:55.169139 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mhz2w"] Dec 06 04:08:55 crc kubenswrapper[4718]: I1206 04:08:55.169272 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:55 crc kubenswrapper[4718]: E1206 04:08:55.169386 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:55 crc kubenswrapper[4718]: I1206 04:08:55.328166 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:55 crc kubenswrapper[4718]: I1206 04:08:55.328219 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:55 crc kubenswrapper[4718]: E1206 04:08:55.328333 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:55 crc kubenswrapper[4718]: I1206 04:08:55.328394 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:55 crc kubenswrapper[4718]: E1206 04:08:55.328549 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:55 crc kubenswrapper[4718]: E1206 04:08:55.328645 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:56 crc kubenswrapper[4718]: E1206 04:08:56.430412 4718 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 04:08:57 crc kubenswrapper[4718]: I1206 04:08:57.328106 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:57 crc kubenswrapper[4718]: I1206 04:08:57.328132 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:57 crc kubenswrapper[4718]: I1206 04:08:57.328173 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:57 crc kubenswrapper[4718]: E1206 04:08:57.328277 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:08:57 crc kubenswrapper[4718]: I1206 04:08:57.328310 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:57 crc kubenswrapper[4718]: E1206 04:08:57.329004 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:57 crc kubenswrapper[4718]: E1206 04:08:57.329373 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:57 crc kubenswrapper[4718]: E1206 04:08:57.329608 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:59 crc kubenswrapper[4718]: I1206 04:08:59.327388 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:08:59 crc kubenswrapper[4718]: I1206 04:08:59.327409 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:08:59 crc kubenswrapper[4718]: I1206 04:08:59.327459 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:08:59 crc kubenswrapper[4718]: E1206 04:08:59.327929 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:08:59 crc kubenswrapper[4718]: E1206 04:08:59.327817 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:08:59 crc kubenswrapper[4718]: E1206 04:08:59.328017 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:08:59 crc kubenswrapper[4718]: I1206 04:08:59.327518 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:08:59 crc kubenswrapper[4718]: E1206 04:08:59.328123 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:09:00 crc kubenswrapper[4718]: I1206 04:09:00.327568 4718 scope.go:117] "RemoveContainer" containerID="c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a" Dec 06 04:09:01 crc kubenswrapper[4718]: I1206 04:09:01.010825 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4r7xc_2272fbce-ce87-4f9f-b1c5-f83f262534de/kube-multus/1.log" Dec 06 04:09:01 crc kubenswrapper[4718]: I1206 04:09:01.010883 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4r7xc" event={"ID":"2272fbce-ce87-4f9f-b1c5-f83f262534de","Type":"ContainerStarted","Data":"5dc497939199c4b153bf5866b448171883409c3dfde61d01d920b1194fc12c52"} Dec 06 04:09:01 crc kubenswrapper[4718]: I1206 04:09:01.328095 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:09:01 crc kubenswrapper[4718]: I1206 04:09:01.328089 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:09:01 crc kubenswrapper[4718]: I1206 04:09:01.328145 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:09:01 crc kubenswrapper[4718]: I1206 04:09:01.328153 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:09:01 crc kubenswrapper[4718]: E1206 04:09:01.329497 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:09:01 crc kubenswrapper[4718]: E1206 04:09:01.329626 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:09:01 crc kubenswrapper[4718]: E1206 04:09:01.329747 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:09:01 crc kubenswrapper[4718]: E1206 04:09:01.329812 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:09:01 crc kubenswrapper[4718]: E1206 04:09:01.431456 4718 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 04:09:03 crc kubenswrapper[4718]: I1206 04:09:03.328405 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:09:03 crc kubenswrapper[4718]: I1206 04:09:03.328395 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:09:03 crc kubenswrapper[4718]: E1206 04:09:03.328604 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:09:03 crc kubenswrapper[4718]: I1206 04:09:03.328422 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:09:03 crc kubenswrapper[4718]: I1206 04:09:03.328410 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:09:03 crc kubenswrapper[4718]: E1206 04:09:03.328772 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:09:03 crc kubenswrapper[4718]: E1206 04:09:03.328924 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:09:03 crc kubenswrapper[4718]: E1206 04:09:03.329049 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:09:05 crc kubenswrapper[4718]: I1206 04:09:05.328326 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:09:05 crc kubenswrapper[4718]: I1206 04:09:05.328354 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:09:05 crc kubenswrapper[4718]: I1206 04:09:05.328371 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:09:05 crc kubenswrapper[4718]: I1206 04:09:05.328326 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:09:05 crc kubenswrapper[4718]: E1206 04:09:05.328490 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 04:09:05 crc kubenswrapper[4718]: E1206 04:09:05.328600 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 04:09:05 crc kubenswrapper[4718]: E1206 04:09:05.328696 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mhz2w" podUID="38892dbb-af9b-4483-bfe6-c1ed160d356c" Dec 06 04:09:05 crc kubenswrapper[4718]: E1206 04:09:05.328820 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.328128 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.328137 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.328300 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.328418 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.331491 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.332347 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.332442 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.332995 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.333405 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 04:09:07 crc kubenswrapper[4718]: I1206 04:09:07.334069 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.064401 4718 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.115071 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.115854 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.117307 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q7kbb"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.118281 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.123315 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7fbd"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.123833 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.124517 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.125210 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.125423 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.126279 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.126532 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.126368 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.126413 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.126924 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6jmn4"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.127805 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.127999 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.128624 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.129072 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.129875 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.131419 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.131603 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.131818 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132087 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132136 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132197 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132369 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132408 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132434 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132564 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132708 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132747 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.132377 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.133014 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.133187 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.136811 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.137402 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.137655 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.137935 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.138220 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.138578 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.139584 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.139901 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.140209 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.141339 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.141940 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb297"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.142446 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.142495 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.145701 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.146197 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.146429 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.148669 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.148910 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.149278 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.149409 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.149844 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.160793 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.161108 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.162300 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.164297 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2zd9s"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.165038 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.166063 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.166216 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.166422 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.169282 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.169913 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170012 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170223 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170453 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-29cng"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170463 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170752 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170859 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170502 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170555 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.171443 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170594 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170625 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.170650 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.171688 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.171790 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.171816 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.172052 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.172104 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.172393 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.172597 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.172697 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-69v98"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.174202 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.177800 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5z699"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.178385 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-v6j6g"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.178653 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.179383 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-f84lw"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.179431 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.179909 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-f84lw" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.180144 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5z699" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.180458 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.180680 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.180809 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.180879 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.180919 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.181101 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.181143 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.181776 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7fbd"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.182294 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.188797 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.192757 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.192954 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193003 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193111 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193304 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q7kbb"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193447 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193470 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7412230b-433d-4a8b-86d7-38cad91e6ecb-encryption-config\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193531 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x8bv\" (UniqueName: \"kubernetes.io/projected/ab65dab6-02da-469a-ba8c-041327284eea-kube-api-access-6x8bv\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193558 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193595 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/79292fb4-57be-4bbe-aa46-09d97f06b109-encryption-config\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193617 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/79292fb4-57be-4bbe-aa46-09d97f06b109-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193652 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ab65dab6-02da-469a-ba8c-041327284eea-machine-approver-tls\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193664 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193669 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7412230b-433d-4a8b-86d7-38cad91e6ecb-audit-dir\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193688 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7412230b-433d-4a8b-86d7-38cad91e6ecb-serving-cert\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193705 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d02801a-a515-4f1f-99a5-90863caec537-config\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193749 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjnh8\" (UniqueName: \"kubernetes.io/projected/5d02801a-a515-4f1f-99a5-90863caec537-kube-api-access-xjnh8\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193768 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-client-ca\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193787 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-etcd-serving-ca\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193833 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab65dab6-02da-469a-ba8c-041327284eea-config\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193866 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7412230b-433d-4a8b-86d7-38cad91e6ecb-etcd-client\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193909 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193926 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ab65dab6-02da-469a-ba8c-041327284eea-auth-proxy-config\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193943 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.193981 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79292fb4-57be-4bbe-aa46-09d97f06b109-serving-cert\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194007 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-image-import-ca\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194160 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-config\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194189 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4kw9\" (UniqueName: \"kubernetes.io/projected/7412230b-433d-4a8b-86d7-38cad91e6ecb-kube-api-access-c4kw9\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194248 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b13c7e13-af0d-4c28-bb48-e457a608075e-serving-cert\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194267 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/79292fb4-57be-4bbe-aa46-09d97f06b109-audit-dir\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194306 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-config\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194322 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/79292fb4-57be-4bbe-aa46-09d97f06b109-etcd-client\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194337 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-audit\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194353 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d02801a-a515-4f1f-99a5-90863caec537-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194393 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/79292fb4-57be-4bbe-aa46-09d97f06b109-audit-policies\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194424 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7412230b-433d-4a8b-86d7-38cad91e6ecb-node-pullsecrets\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194440 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5d02801a-a515-4f1f-99a5-90863caec537-images\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194484 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/79292fb4-57be-4bbe-aa46-09d97f06b109-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194500 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8m87\" (UniqueName: \"kubernetes.io/projected/b13c7e13-af0d-4c28-bb48-e457a608075e-kube-api-access-l8m87\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194546 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ddpk\" (UniqueName: \"kubernetes.io/projected/79292fb4-57be-4bbe-aa46-09d97f06b109-kube-api-access-2ddpk\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.194089 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.195435 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.195729 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.195927 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.196107 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.196306 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.196511 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.197245 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.197414 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.197571 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.197710 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.197943 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.198776 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.198838 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.199028 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.199108 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.199195 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.199463 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.199529 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.211709 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.212149 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.212198 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.213001 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6jmn4"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.217130 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5z699"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.219825 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.221888 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.225341 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.225628 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.228393 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.233502 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb297"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.233542 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2zd9s"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.233552 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.234487 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.235519 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-v6j6g"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.235977 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-f84lw"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.237125 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-69v98"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.240897 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.240957 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-29cng"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.241536 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.266409 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.285747 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-x2dzh"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.287209 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.287514 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.287615 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.288089 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.288109 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.288989 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sghf9"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.289467 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.289596 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.289861 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.289865 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.290715 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.291345 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.292384 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.292979 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.293592 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.293594 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.295587 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7412230b-433d-4a8b-86d7-38cad91e6ecb-encryption-config\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.295662 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x8bv\" (UniqueName: \"kubernetes.io/projected/ab65dab6-02da-469a-ba8c-041327284eea-kube-api-access-6x8bv\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.295735 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77330fe5-fc23-41b0-9b9d-e6497908bd77-service-ca-bundle\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.295763 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-client-ca\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.295826 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-oauth-serving-cert\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.295852 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c76e70f2-707c-4d55-8af2-672fac1f7ac4-serving-cert\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.295915 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/79292fb4-57be-4bbe-aa46-09d97f06b109-encryption-config\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.295939 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqmfw\" (UniqueName: \"kubernetes.io/projected/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-kube-api-access-kqmfw\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296008 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lzvl8\" (UID: \"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296031 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-trusted-ca-bundle\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296057 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/79292fb4-57be-4bbe-aa46-09d97f06b109-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296113 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ab65dab6-02da-469a-ba8c-041327284eea-machine-approver-tls\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296138 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77330fe5-fc23-41b0-9b9d-e6497908bd77-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296213 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4ffee8e-d1ea-4908-b160-473b1a73d31c-serving-cert\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296310 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxf8j\" (UniqueName: \"kubernetes.io/projected/58ce50cf-28a9-43a5-b4b3-bee39bbf7393-kube-api-access-dxf8j\") pod \"openshift-config-operator-7777fb866f-29cng\" (UID: \"58ce50cf-28a9-43a5-b4b3-bee39bbf7393\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296340 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7412230b-433d-4a8b-86d7-38cad91e6ecb-audit-dir\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296404 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/58ce50cf-28a9-43a5-b4b3-bee39bbf7393-available-featuregates\") pod \"openshift-config-operator-7777fb866f-29cng\" (UID: \"58ce50cf-28a9-43a5-b4b3-bee39bbf7393\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296431 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296453 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296517 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296579 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77330fe5-fc23-41b0-9b9d-e6497908bd77-config\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296601 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbc5x\" (UniqueName: \"kubernetes.io/projected/77330fe5-fc23-41b0-9b9d-e6497908bd77-kube-api-access-xbc5x\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296641 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296666 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7412230b-433d-4a8b-86d7-38cad91e6ecb-serving-cert\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296689 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr59d\" (UniqueName: \"kubernetes.io/projected/1308cc0b-85c6-49e9-a036-fc110f461f9c-kube-api-access-nr59d\") pod \"downloads-7954f5f757-f84lw\" (UID: \"1308cc0b-85c6-49e9-a036-fc110f461f9c\") " pod="openshift-console/downloads-7954f5f757-f84lw" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296712 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4ffee8e-d1ea-4908-b160-473b1a73d31c-config\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296790 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d02801a-a515-4f1f-99a5-90863caec537-config\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296814 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjnh8\" (UniqueName: \"kubernetes.io/projected/5d02801a-a515-4f1f-99a5-90863caec537-kube-api-access-xjnh8\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296840 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-dir\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296865 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77ds4\" (UniqueName: \"kubernetes.io/projected/83164f34-cd50-4f7d-b6ae-f9e65d3b69fe-kube-api-access-77ds4\") pod \"openshift-apiserver-operator-796bbdcf4f-74rb2\" (UID: \"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.296891 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lzvl8\" (UID: \"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.297111 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-client-ca\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.297141 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/454877ff-cab8-4340-bfe7-1709bbd1227d-console-serving-cert\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.297200 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7412230b-433d-4a8b-86d7-38cad91e6ecb-audit-dir\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.297164 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e812f183-3419-4365-99f5-94f1e9f8abcb-metrics-tls\") pod \"dns-operator-744455d44c-5z699\" (UID: \"e812f183-3419-4365-99f5-94f1e9f8abcb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5z699" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.297386 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlgqt\" (UniqueName: \"kubernetes.io/projected/c76e70f2-707c-4d55-8af2-672fac1f7ac4-kube-api-access-mlgqt\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.297821 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-etcd-serving-ca\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.297860 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab65dab6-02da-469a-ba8c-041327284eea-config\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.298213 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-service-ca\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.298422 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83164f34-cd50-4f7d-b6ae-f9e65d3b69fe-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-74rb2\" (UID: \"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.298447 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvbq9\" (UniqueName: \"kubernetes.io/projected/e812f183-3419-4365-99f5-94f1e9f8abcb-kube-api-access-lvbq9\") pod \"dns-operator-744455d44c-5z699\" (UID: \"e812f183-3419-4365-99f5-94f1e9f8abcb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5z699" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.299019 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-client-ca\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.299765 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7412230b-433d-4a8b-86d7-38cad91e6ecb-etcd-client\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301476 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab65dab6-02da-469a-ba8c-041327284eea-config\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301521 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301543 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-etcd-serving-ca\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301562 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301615 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjrqq\" (UniqueName: \"kubernetes.io/projected/ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5-kube-api-access-qjrqq\") pod \"openshift-controller-manager-operator-756b6f6bc6-lzvl8\" (UID: \"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301638 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckrfk\" (UniqueName: \"kubernetes.io/projected/7ef554fa-0472-493f-b4ac-c131e53eb901-kube-api-access-ckrfk\") pod \"cluster-samples-operator-665b6dd947-kwhws\" (UID: \"7ef554fa-0472-493f-b4ac-c131e53eb901\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301669 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58ce50cf-28a9-43a5-b4b3-bee39bbf7393-serving-cert\") pod \"openshift-config-operator-7777fb866f-29cng\" (UID: \"58ce50cf-28a9-43a5-b4b3-bee39bbf7393\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301714 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301769 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ab65dab6-02da-469a-ba8c-041327284eea-auth-proxy-config\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.301977 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302011 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/454877ff-cab8-4340-bfe7-1709bbd1227d-console-oauth-config\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302054 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302087 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79292fb4-57be-4bbe-aa46-09d97f06b109-serving-cert\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302103 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302121 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-image-import-ca\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302138 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4ffee8e-d1ea-4908-b160-473b1a73d31c-trusted-ca\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302153 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz92z\" (UniqueName: \"kubernetes.io/projected/f4ffee8e-d1ea-4908-b160-473b1a73d31c-kube-api-access-pz92z\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302167 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-config\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302186 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4kw9\" (UniqueName: \"kubernetes.io/projected/7412230b-433d-4a8b-86d7-38cad91e6ecb-kube-api-access-c4kw9\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302201 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-config\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302219 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77330fe5-fc23-41b0-9b9d-e6497908bd77-serving-cert\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302285 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ab65dab6-02da-469a-ba8c-041327284eea-auth-proxy-config\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302301 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83164f34-cd50-4f7d-b6ae-f9e65d3b69fe-config\") pod \"openshift-apiserver-operator-796bbdcf4f-74rb2\" (UID: \"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302349 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b13c7e13-af0d-4c28-bb48-e457a608075e-serving-cert\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302369 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-config\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302387 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/79292fb4-57be-4bbe-aa46-09d97f06b109-etcd-client\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302403 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/79292fb4-57be-4bbe-aa46-09d97f06b109-audit-dir\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302458 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-audit\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302502 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d02801a-a515-4f1f-99a5-90863caec537-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302520 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302536 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302558 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/79292fb4-57be-4bbe-aa46-09d97f06b109-audit-policies\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302576 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7412230b-433d-4a8b-86d7-38cad91e6ecb-node-pullsecrets\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302596 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-policies\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302620 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5d02801a-a515-4f1f-99a5-90863caec537-images\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302637 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d02801a-a515-4f1f-99a5-90863caec537-config\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302676 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/79292fb4-57be-4bbe-aa46-09d97f06b109-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302730 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/79292fb4-57be-4bbe-aa46-09d97f06b109-audit-dir\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.302643 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/79292fb4-57be-4bbe-aa46-09d97f06b109-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.303059 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/79292fb4-57be-4bbe-aa46-09d97f06b109-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.303117 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8m87\" (UniqueName: \"kubernetes.io/projected/b13c7e13-af0d-4c28-bb48-e457a608075e-kube-api-access-l8m87\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.303165 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ddpk\" (UniqueName: \"kubernetes.io/projected/79292fb4-57be-4bbe-aa46-09d97f06b109-kube-api-access-2ddpk\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.303573 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-console-config\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.303608 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.303632 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m97vj\" (UniqueName: \"kubernetes.io/projected/454877ff-cab8-4340-bfe7-1709bbd1227d-kube-api-access-m97vj\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.303657 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ef554fa-0472-493f-b4ac-c131e53eb901-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-kwhws\" (UID: \"7ef554fa-0472-493f-b4ac-c131e53eb901\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.303656 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/79292fb4-57be-4bbe-aa46-09d97f06b109-encryption-config\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.303872 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ab65dab6-02da-469a-ba8c-041327284eea-machine-approver-tls\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.304094 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7412230b-433d-4a8b-86d7-38cad91e6ecb-node-pullsecrets\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.304359 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/79292fb4-57be-4bbe-aa46-09d97f06b109-audit-policies\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.304823 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5d02801a-a515-4f1f-99a5-90863caec537-images\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.304859 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.305574 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.305923 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-w4fdl"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.307262 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.312152 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-7jwsr"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.312364 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.313104 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d02801a-a515-4f1f-99a5-90863caec537-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.313216 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.313515 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.313320 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.313885 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7412230b-433d-4a8b-86d7-38cad91e6ecb-encryption-config\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.313966 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79292fb4-57be-4bbe-aa46-09d97f06b109-serving-cert\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.314172 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.314560 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-config\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.314811 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-audit\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.314829 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7412230b-433d-4a8b-86d7-38cad91e6ecb-image-import-ca\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.314963 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.315315 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-x2dzh"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.315334 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vp8rk"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.315841 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7412230b-433d-4a8b-86d7-38cad91e6ecb-serving-cert\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.315869 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.315898 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.318874 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b13c7e13-af0d-4c28-bb48-e457a608075e-serving-cert\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.324612 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7412230b-433d-4a8b-86d7-38cad91e6ecb-etcd-client\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.328186 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/79292fb4-57be-4bbe-aa46-09d97f06b109-etcd-client\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.329414 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-config\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.339296 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.342044 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.343041 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.350629 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.352208 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.352315 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.364896 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.366379 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.368163 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.371061 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.374656 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.381202 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.381386 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.386196 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6kwkv"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.386982 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.387827 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vqswv"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.388212 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.388831 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-d96mp"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.389748 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.389844 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.390594 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.390974 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gfknw"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.391925 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.391928 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.392895 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.393305 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7jwsr"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.395064 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vp8rk"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.397962 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.399513 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sghf9"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.404060 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405184 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c76e70f2-707c-4d55-8af2-672fac1f7ac4-serving-cert\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405224 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqmfw\" (UniqueName: \"kubernetes.io/projected/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-kube-api-access-kqmfw\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405263 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lzvl8\" (UID: \"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405282 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-trusted-ca-bundle\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405302 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77330fe5-fc23-41b0-9b9d-e6497908bd77-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405394 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4ffee8e-d1ea-4908-b160-473b1a73d31c-serving-cert\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405415 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxf8j\" (UniqueName: \"kubernetes.io/projected/58ce50cf-28a9-43a5-b4b3-bee39bbf7393-kube-api-access-dxf8j\") pod \"openshift-config-operator-7777fb866f-29cng\" (UID: \"58ce50cf-28a9-43a5-b4b3-bee39bbf7393\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405437 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/58ce50cf-28a9-43a5-b4b3-bee39bbf7393-available-featuregates\") pod \"openshift-config-operator-7777fb866f-29cng\" (UID: \"58ce50cf-28a9-43a5-b4b3-bee39bbf7393\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405454 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405473 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405491 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405508 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77330fe5-fc23-41b0-9b9d-e6497908bd77-config\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405524 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbc5x\" (UniqueName: \"kubernetes.io/projected/77330fe5-fc23-41b0-9b9d-e6497908bd77-kube-api-access-xbc5x\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405541 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405563 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89k9s\" (UniqueName: \"kubernetes.io/projected/8204ea9d-c0c7-4541-8e4e-a1b96303aa54-kube-api-access-89k9s\") pod \"migrator-59844c95c7-c4lmg\" (UID: \"8204ea9d-c0c7-4541-8e4e-a1b96303aa54\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405584 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b619ef6-1db6-4a92-9810-aa00e60f5bb8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5kg75\" (UID: \"8b619ef6-1db6-4a92-9810-aa00e60f5bb8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405601 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr59d\" (UniqueName: \"kubernetes.io/projected/1308cc0b-85c6-49e9-a036-fc110f461f9c-kube-api-access-nr59d\") pod \"downloads-7954f5f757-f84lw\" (UID: \"1308cc0b-85c6-49e9-a036-fc110f461f9c\") " pod="openshift-console/downloads-7954f5f757-f84lw" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405625 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4ffee8e-d1ea-4908-b160-473b1a73d31c-config\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405648 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-dir\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405665 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77ds4\" (UniqueName: \"kubernetes.io/projected/83164f34-cd50-4f7d-b6ae-f9e65d3b69fe-kube-api-access-77ds4\") pod \"openshift-apiserver-operator-796bbdcf4f-74rb2\" (UID: \"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405682 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lzvl8\" (UID: \"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405700 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cbccb659-e964-4146-a586-d71b85f000f0-trusted-ca\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405716 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/454877ff-cab8-4340-bfe7-1709bbd1227d-console-serving-cert\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405754 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e812f183-3419-4365-99f5-94f1e9f8abcb-metrics-tls\") pod \"dns-operator-744455d44c-5z699\" (UID: \"e812f183-3419-4365-99f5-94f1e9f8abcb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5z699" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405770 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlgqt\" (UniqueName: \"kubernetes.io/projected/c76e70f2-707c-4d55-8af2-672fac1f7ac4-kube-api-access-mlgqt\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405793 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b619ef6-1db6-4a92-9810-aa00e60f5bb8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5kg75\" (UID: \"8b619ef6-1db6-4a92-9810-aa00e60f5bb8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405810 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-service-ca\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405827 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/64767824-fed2-496b-baa5-2c3ce0f8ee5b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405854 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83164f34-cd50-4f7d-b6ae-f9e65d3b69fe-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-74rb2\" (UID: \"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405871 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvbq9\" (UniqueName: \"kubernetes.io/projected/e812f183-3419-4365-99f5-94f1e9f8abcb-kube-api-access-lvbq9\") pod \"dns-operator-744455d44c-5z699\" (UID: \"e812f183-3419-4365-99f5-94f1e9f8abcb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5z699" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405889 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/96226f3a-94e6-4895-a9fc-662f07f8cdf8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qsr\" (UID: \"96226f3a-94e6-4895-a9fc-662f07f8cdf8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405915 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405932 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.405948 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjrqq\" (UniqueName: \"kubernetes.io/projected/ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5-kube-api-access-qjrqq\") pod \"openshift-controller-manager-operator-756b6f6bc6-lzvl8\" (UID: \"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406043 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckrfk\" (UniqueName: \"kubernetes.io/projected/7ef554fa-0472-493f-b4ac-c131e53eb901-kube-api-access-ckrfk\") pod \"cluster-samples-operator-665b6dd947-kwhws\" (UID: \"7ef554fa-0472-493f-b4ac-c131e53eb901\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406066 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58ce50cf-28a9-43a5-b4b3-bee39bbf7393-serving-cert\") pod \"openshift-config-operator-7777fb866f-29cng\" (UID: \"58ce50cf-28a9-43a5-b4b3-bee39bbf7393\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406090 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406107 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/454877ff-cab8-4340-bfe7-1709bbd1227d-console-oauth-config\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406125 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfc4d\" (UniqueName: \"kubernetes.io/projected/daeca836-1e78-4156-9cbe-b065c3d2f59f-kube-api-access-qfc4d\") pod \"multus-admission-controller-857f4d67dd-vp8rk\" (UID: \"daeca836-1e78-4156-9cbe-b065c3d2f59f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406145 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406162 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4ffee8e-d1ea-4908-b160-473b1a73d31c-trusted-ca\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406183 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz92z\" (UniqueName: \"kubernetes.io/projected/f4ffee8e-d1ea-4908-b160-473b1a73d31c-kube-api-access-pz92z\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406201 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-config\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406226 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77330fe5-fc23-41b0-9b9d-e6497908bd77-serving-cert\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406263 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83164f34-cd50-4f7d-b6ae-f9e65d3b69fe-config\") pod \"openshift-apiserver-operator-796bbdcf4f-74rb2\" (UID: \"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406279 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78pds\" (UniqueName: \"kubernetes.io/projected/64767824-fed2-496b-baa5-2c3ce0f8ee5b-kube-api-access-78pds\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406296 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cbccb659-e964-4146-a586-d71b85f000f0-metrics-tls\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406316 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cbccb659-e964-4146-a586-d71b85f000f0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406335 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/64767824-fed2-496b-baa5-2c3ce0f8ee5b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406354 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406370 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406469 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-policies\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406487 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/64767824-fed2-496b-baa5-2c3ce0f8ee5b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406494 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406505 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xknn\" (UniqueName: \"kubernetes.io/projected/96226f3a-94e6-4895-a9fc-662f07f8cdf8-kube-api-access-4xknn\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qsr\" (UID: \"96226f3a-94e6-4895-a9fc-662f07f8cdf8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406527 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/daeca836-1e78-4156-9cbe-b065c3d2f59f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vp8rk\" (UID: \"daeca836-1e78-4156-9cbe-b065c3d2f59f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406543 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87px2\" (UniqueName: \"kubernetes.io/projected/cbccb659-e964-4146-a586-d71b85f000f0-kube-api-access-87px2\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406577 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-console-config\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406599 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406616 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m97vj\" (UniqueName: \"kubernetes.io/projected/454877ff-cab8-4340-bfe7-1709bbd1227d-kube-api-access-m97vj\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406641 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ef554fa-0472-493f-b4ac-c131e53eb901-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-kwhws\" (UID: \"7ef554fa-0472-493f-b4ac-c131e53eb901\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406657 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2fhc\" (UniqueName: \"kubernetes.io/projected/8b619ef6-1db6-4a92-9810-aa00e60f5bb8-kube-api-access-r2fhc\") pod \"kube-storage-version-migrator-operator-b67b599dd-5kg75\" (UID: \"8b619ef6-1db6-4a92-9810-aa00e60f5bb8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406703 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77330fe5-fc23-41b0-9b9d-e6497908bd77-service-ca-bundle\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406720 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-client-ca\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406739 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-oauth-serving-cert\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.406842 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77330fe5-fc23-41b0-9b9d-e6497908bd77-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.407078 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-dir\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.407158 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/58ce50cf-28a9-43a5-b4b3-bee39bbf7393-available-featuregates\") pod \"openshift-config-operator-7777fb866f-29cng\" (UID: \"58ce50cf-28a9-43a5-b4b3-bee39bbf7393\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.407288 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-trusted-ca-bundle\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.407826 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.408075 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lzvl8\" (UID: \"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.408539 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77330fe5-fc23-41b0-9b9d-e6497908bd77-config\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.408553 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-console-config\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.410872 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.411141 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.412009 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-policies\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.412345 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-oauth-serving-cert\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.412635 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-client-ca\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.412858 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4ffee8e-d1ea-4908-b160-473b1a73d31c-serving-cert\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.413270 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/454877ff-cab8-4340-bfe7-1709bbd1227d-service-ca\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.413316 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4ffee8e-d1ea-4908-b160-473b1a73d31c-config\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.413507 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c76e70f2-707c-4d55-8af2-672fac1f7ac4-serving-cert\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.414118 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-config\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.414283 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lzvl8\" (UID: \"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.414590 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.415194 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77330fe5-fc23-41b0-9b9d-e6497908bd77-service-ca-bundle\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.415304 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83164f34-cd50-4f7d-b6ae-f9e65d3b69fe-config\") pod \"openshift-apiserver-operator-796bbdcf4f-74rb2\" (UID: \"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.415388 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4ffee8e-d1ea-4908-b160-473b1a73d31c-trusted-ca\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.416065 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.416617 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77330fe5-fc23-41b0-9b9d-e6497908bd77-serving-cert\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.416837 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.417639 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e812f183-3419-4365-99f5-94f1e9f8abcb-metrics-tls\") pod \"dns-operator-744455d44c-5z699\" (UID: \"e812f183-3419-4365-99f5-94f1e9f8abcb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5z699" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.417773 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.418070 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83164f34-cd50-4f7d-b6ae-f9e65d3b69fe-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-74rb2\" (UID: \"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.418084 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.418116 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/454877ff-cab8-4340-bfe7-1709bbd1227d-console-serving-cert\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.418141 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/454877ff-cab8-4340-bfe7-1709bbd1227d-console-oauth-config\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.418386 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.418418 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.419181 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.419202 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.419271 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ef554fa-0472-493f-b4ac-c131e53eb901-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-kwhws\" (UID: \"7ef554fa-0472-493f-b4ac-c131e53eb901\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.419971 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58ce50cf-28a9-43a5-b4b3-bee39bbf7393-serving-cert\") pod \"openshift-config-operator-7777fb866f-29cng\" (UID: \"58ce50cf-28a9-43a5-b4b3-bee39bbf7393\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.420063 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-mxjzh"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.421480 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mxjzh" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.422614 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.422735 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.423337 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.424042 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.425078 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.425433 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.430164 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.430200 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-59sd7"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.431160 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6kwkv"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.431266 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.434286 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gfknw"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.434869 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.436126 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.437319 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vqswv"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.438621 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.439613 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.439726 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.441807 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mxjzh"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.443317 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.444263 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-59sd7"] Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.458668 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.477977 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.498138 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.507785 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/64767824-fed2-496b-baa5-2c3ce0f8ee5b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.507850 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xknn\" (UniqueName: \"kubernetes.io/projected/96226f3a-94e6-4895-a9fc-662f07f8cdf8-kube-api-access-4xknn\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qsr\" (UID: \"96226f3a-94e6-4895-a9fc-662f07f8cdf8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.507872 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87px2\" (UniqueName: \"kubernetes.io/projected/cbccb659-e964-4146-a586-d71b85f000f0-kube-api-access-87px2\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.507918 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/daeca836-1e78-4156-9cbe-b065c3d2f59f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vp8rk\" (UID: \"daeca836-1e78-4156-9cbe-b065c3d2f59f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.507953 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2fhc\" (UniqueName: \"kubernetes.io/projected/8b619ef6-1db6-4a92-9810-aa00e60f5bb8-kube-api-access-r2fhc\") pod \"kube-storage-version-migrator-operator-b67b599dd-5kg75\" (UID: \"8b619ef6-1db6-4a92-9810-aa00e60f5bb8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508083 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89k9s\" (UniqueName: \"kubernetes.io/projected/8204ea9d-c0c7-4541-8e4e-a1b96303aa54-kube-api-access-89k9s\") pod \"migrator-59844c95c7-c4lmg\" (UID: \"8204ea9d-c0c7-4541-8e4e-a1b96303aa54\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508113 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b619ef6-1db6-4a92-9810-aa00e60f5bb8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5kg75\" (UID: \"8b619ef6-1db6-4a92-9810-aa00e60f5bb8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508190 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cbccb659-e964-4146-a586-d71b85f000f0-trusted-ca\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508219 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b619ef6-1db6-4a92-9810-aa00e60f5bb8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5kg75\" (UID: \"8b619ef6-1db6-4a92-9810-aa00e60f5bb8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508251 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/64767824-fed2-496b-baa5-2c3ce0f8ee5b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508286 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/96226f3a-94e6-4895-a9fc-662f07f8cdf8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qsr\" (UID: \"96226f3a-94e6-4895-a9fc-662f07f8cdf8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508361 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfc4d\" (UniqueName: \"kubernetes.io/projected/daeca836-1e78-4156-9cbe-b065c3d2f59f-kube-api-access-qfc4d\") pod \"multus-admission-controller-857f4d67dd-vp8rk\" (UID: \"daeca836-1e78-4156-9cbe-b065c3d2f59f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508399 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78pds\" (UniqueName: \"kubernetes.io/projected/64767824-fed2-496b-baa5-2c3ce0f8ee5b-kube-api-access-78pds\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508427 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cbccb659-e964-4146-a586-d71b85f000f0-metrics-tls\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508449 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cbccb659-e964-4146-a586-d71b85f000f0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.508474 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/64767824-fed2-496b-baa5-2c3ce0f8ee5b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.518567 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.521976 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cbccb659-e964-4146-a586-d71b85f000f0-metrics-tls\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.538166 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.569295 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.578434 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.579409 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cbccb659-e964-4146-a586-d71b85f000f0-trusted-ca\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.598824 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.618788 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.637856 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.667693 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.669940 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/64767824-fed2-496b-baa5-2c3ce0f8ee5b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.678185 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.698806 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.718578 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.738980 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.758840 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.779880 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.793134 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b619ef6-1db6-4a92-9810-aa00e60f5bb8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5kg75\" (UID: \"8b619ef6-1db6-4a92-9810-aa00e60f5bb8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.799523 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.818677 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.829524 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b619ef6-1db6-4a92-9810-aa00e60f5bb8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5kg75\" (UID: \"8b619ef6-1db6-4a92-9810-aa00e60f5bb8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.831349 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.839891 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.859740 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.871457 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/64767824-fed2-496b-baa5-2c3ce0f8ee5b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.879058 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.898672 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.918939 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.939214 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.958956 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.979405 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 04:09:15 crc kubenswrapper[4718]: I1206 04:09:15.998587 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.036651 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjnh8\" (UniqueName: \"kubernetes.io/projected/5d02801a-a515-4f1f-99a5-90863caec537-kube-api-access-xjnh8\") pod \"machine-api-operator-5694c8668f-6jmn4\" (UID: \"5d02801a-a515-4f1f-99a5-90863caec537\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.052854 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x8bv\" (UniqueName: \"kubernetes.io/projected/ab65dab6-02da-469a-ba8c-041327284eea-kube-api-access-6x8bv\") pod \"machine-approver-56656f9798-xmfbd\" (UID: \"ab65dab6-02da-469a-ba8c-041327284eea\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.067133 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.077136 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4kw9\" (UniqueName: \"kubernetes.io/projected/7412230b-433d-4a8b-86d7-38cad91e6ecb-kube-api-access-c4kw9\") pod \"apiserver-76f77b778f-q7kbb\" (UID: \"7412230b-433d-4a8b-86d7-38cad91e6ecb\") " pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.085381 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.092905 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ddpk\" (UniqueName: \"kubernetes.io/projected/79292fb4-57be-4bbe-aa46-09d97f06b109-kube-api-access-2ddpk\") pod \"apiserver-7bbb656c7d-krdhn\" (UID: \"79292fb4-57be-4bbe-aa46-09d97f06b109\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.118340 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.121440 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8m87\" (UniqueName: \"kubernetes.io/projected/b13c7e13-af0d-4c28-bb48-e457a608075e-kube-api-access-l8m87\") pod \"controller-manager-879f6c89f-v7fbd\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.128478 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.139135 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.139538 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.153159 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/96226f3a-94e6-4895-a9fc-662f07f8cdf8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qsr\" (UID: \"96226f3a-94e6-4895-a9fc-662f07f8cdf8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.159078 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.181780 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.199675 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.219785 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.239387 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.260215 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.278536 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.301177 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.302642 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q7kbb"] Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.317140 4718 request.go:700] Waited for 1.003002606s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0 Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.319383 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.341967 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.349768 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn"] Dec 06 04:09:16 crc kubenswrapper[4718]: W1206 04:09:16.353435 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79292fb4_57be_4bbe_aa46_09d97f06b109.slice/crio-6bc83a9dd9f44dab5c46cf988465f513cd0c6e682d9f283b6540937874aeaef8 WatchSource:0}: Error finding container 6bc83a9dd9f44dab5c46cf988465f513cd0c6e682d9f283b6540937874aeaef8: Status 404 returned error can't find the container with id 6bc83a9dd9f44dab5c46cf988465f513cd0c6e682d9f283b6540937874aeaef8 Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.358868 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.380490 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.381423 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6jmn4"] Dec 06 04:09:16 crc kubenswrapper[4718]: W1206 04:09:16.398559 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d02801a_a515_4f1f_99a5_90863caec537.slice/crio-f468c148f55fd3649238b878d0b5abe9e34e458b350e3b142e9d11093729a2be WatchSource:0}: Error finding container f468c148f55fd3649238b878d0b5abe9e34e458b350e3b142e9d11093729a2be: Status 404 returned error can't find the container with id f468c148f55fd3649238b878d0b5abe9e34e458b350e3b142e9d11093729a2be Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.398679 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.399103 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.413910 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/daeca836-1e78-4156-9cbe-b065c3d2f59f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vp8rk\" (UID: \"daeca836-1e78-4156-9cbe-b065c3d2f59f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.418404 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.438962 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.458745 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.480768 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.499688 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.572377 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.578519 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.598715 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.618155 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.627351 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7fbd"] Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.638015 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.658095 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.679452 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.698639 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.718673 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 04:09:16 crc kubenswrapper[4718]: W1206 04:09:16.740837 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb13c7e13_af0d_4c28_bb48_e457a608075e.slice/crio-d61de62e8aa3c22f6fa2ab09211b6b1e1e6033532f37fbb5fec14c346bd0386c WatchSource:0}: Error finding container d61de62e8aa3c22f6fa2ab09211b6b1e1e6033532f37fbb5fec14c346bd0386c: Status 404 returned error can't find the container with id d61de62e8aa3c22f6fa2ab09211b6b1e1e6033532f37fbb5fec14c346bd0386c Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.741065 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.766622 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.777976 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.799763 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.819287 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.839307 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.859711 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.879966 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.900072 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.919021 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.938787 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.959560 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.978794 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 04:09:16 crc kubenswrapper[4718]: I1206 04:09:16.999778 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.018208 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.038917 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.059786 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.080336 4718 generic.go:334] "Generic (PLEG): container finished" podID="79292fb4-57be-4bbe-aa46-09d97f06b109" containerID="69310ff61b276cda2b539226fed1888a725f312249fd943cd6de4368a58a77ad" exitCode=0 Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.080445 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.080478 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" event={"ID":"79292fb4-57be-4bbe-aa46-09d97f06b109","Type":"ContainerDied","Data":"69310ff61b276cda2b539226fed1888a725f312249fd943cd6de4368a58a77ad"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.080522 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" event={"ID":"79292fb4-57be-4bbe-aa46-09d97f06b109","Type":"ContainerStarted","Data":"6bc83a9dd9f44dab5c46cf988465f513cd0c6e682d9f283b6540937874aeaef8"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.082287 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" event={"ID":"5d02801a-a515-4f1f-99a5-90863caec537","Type":"ContainerStarted","Data":"8bf48683e4e11d3c9265cc2c82fd7e6e9796f7dfcde218a4a27bc83da3af419d"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.082364 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" event={"ID":"5d02801a-a515-4f1f-99a5-90863caec537","Type":"ContainerStarted","Data":"be07c896c5121a648dbfc5b3669633b4ba97a3294b5efe509ffcd70e6fa22599"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.082383 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" event={"ID":"5d02801a-a515-4f1f-99a5-90863caec537","Type":"ContainerStarted","Data":"f468c148f55fd3649238b878d0b5abe9e34e458b350e3b142e9d11093729a2be"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.084483 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" event={"ID":"ab65dab6-02da-469a-ba8c-041327284eea","Type":"ContainerStarted","Data":"ba19298ab3c5ff4b174c9eca06f049f9841f2fadb057f28340cb009b420521a1"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.084548 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" event={"ID":"ab65dab6-02da-469a-ba8c-041327284eea","Type":"ContainerStarted","Data":"2dace2257cc8f388265151278ea695f5ab7ed99f5c4eef0332c1066f3c066dad"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.084568 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" event={"ID":"ab65dab6-02da-469a-ba8c-041327284eea","Type":"ContainerStarted","Data":"ff168816237510695cb0ad6720b168117972dd722d5f6f97abf4b97828cbc78a"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.086419 4718 generic.go:334] "Generic (PLEG): container finished" podID="7412230b-433d-4a8b-86d7-38cad91e6ecb" containerID="a51cdcf298cc982c4253af4e66613ddc3617c929c6372cf1fcfe75fd1ef242ee" exitCode=0 Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.086499 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" event={"ID":"7412230b-433d-4a8b-86d7-38cad91e6ecb","Type":"ContainerDied","Data":"a51cdcf298cc982c4253af4e66613ddc3617c929c6372cf1fcfe75fd1ef242ee"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.086528 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" event={"ID":"7412230b-433d-4a8b-86d7-38cad91e6ecb","Type":"ContainerStarted","Data":"d728e35d18063d8e5c730d8b710c76d392d10e42b8a689e9fd096c50d3ccc219"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.088848 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" event={"ID":"b13c7e13-af0d-4c28-bb48-e457a608075e","Type":"ContainerStarted","Data":"04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.088894 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" event={"ID":"b13c7e13-af0d-4c28-bb48-e457a608075e","Type":"ContainerStarted","Data":"d61de62e8aa3c22f6fa2ab09211b6b1e1e6033532f37fbb5fec14c346bd0386c"} Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.089121 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.091455 4718 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-v7fbd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.091503 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" podUID="b13c7e13-af0d-4c28-bb48-e457a608075e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.099997 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.118703 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.158689 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.162318 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqmfw\" (UniqueName: \"kubernetes.io/projected/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-kube-api-access-kqmfw\") pod \"oauth-openshift-558db77b4-pb297\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.194352 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr59d\" (UniqueName: \"kubernetes.io/projected/1308cc0b-85c6-49e9-a036-fc110f461f9c-kube-api-access-nr59d\") pod \"downloads-7954f5f757-f84lw\" (UID: \"1308cc0b-85c6-49e9-a036-fc110f461f9c\") " pod="openshift-console/downloads-7954f5f757-f84lw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.215224 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxf8j\" (UniqueName: \"kubernetes.io/projected/58ce50cf-28a9-43a5-b4b3-bee39bbf7393-kube-api-access-dxf8j\") pod \"openshift-config-operator-7777fb866f-29cng\" (UID: \"58ce50cf-28a9-43a5-b4b3-bee39bbf7393\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.230544 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.230797 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.231038 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:09:17 crc kubenswrapper[4718]: E1206 04:09:17.231905 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:11:19.23188261 +0000 UTC m=+268.237587781 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.232013 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.233566 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77ds4\" (UniqueName: \"kubernetes.io/projected/83164f34-cd50-4f7d-b6ae-f9e65d3b69fe-kube-api-access-77ds4\") pod \"openshift-apiserver-operator-796bbdcf4f-74rb2\" (UID: \"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.234176 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.258086 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbc5x\" (UniqueName: \"kubernetes.io/projected/77330fe5-fc23-41b0-9b9d-e6497908bd77-kube-api-access-xbc5x\") pod \"authentication-operator-69f744f599-2zd9s\" (UID: \"77330fe5-fc23-41b0-9b9d-e6497908bd77\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.269559 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.288915 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlgqt\" (UniqueName: \"kubernetes.io/projected/c76e70f2-707c-4d55-8af2-672fac1f7ac4-kube-api-access-mlgqt\") pod \"route-controller-manager-6576b87f9c-cwfkx\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.298137 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvbq9\" (UniqueName: \"kubernetes.io/projected/e812f183-3419-4365-99f5-94f1e9f8abcb-kube-api-access-lvbq9\") pod \"dns-operator-744455d44c-5z699\" (UID: \"e812f183-3419-4365-99f5-94f1e9f8abcb\") " pod="openshift-dns-operator/dns-operator-744455d44c-5z699" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.317551 4718 request.go:700] Waited for 1.904835444s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-samples-operator/serviceaccounts/cluster-samples-operator/token Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.321577 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjrqq\" (UniqueName: \"kubernetes.io/projected/ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5-kube-api-access-qjrqq\") pod \"openshift-controller-manager-operator-756b6f6bc6-lzvl8\" (UID: \"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.332373 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.332426 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.332972 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckrfk\" (UniqueName: \"kubernetes.io/projected/7ef554fa-0472-493f-b4ac-c131e53eb901-kube-api-access-ckrfk\") pod \"cluster-samples-operator-665b6dd947-kwhws\" (UID: \"7ef554fa-0472-493f-b4ac-c131e53eb901\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.342088 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.348722 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.350393 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.357483 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.364317 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.366621 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.371996 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m97vj\" (UniqueName: \"kubernetes.io/projected/454877ff-cab8-4340-bfe7-1709bbd1227d-kube-api-access-m97vj\") pod \"console-f9d7485db-v6j6g\" (UID: \"454877ff-cab8-4340-bfe7-1709bbd1227d\") " pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.382608 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.393085 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.398592 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.399391 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz92z\" (UniqueName: \"kubernetes.io/projected/f4ffee8e-d1ea-4908-b160-473b1a73d31c-kube-api-access-pz92z\") pod \"console-operator-58897d9998-69v98\" (UID: \"f4ffee8e-d1ea-4908-b160-473b1a73d31c\") " pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.404610 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.419588 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.442384 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.453042 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.459448 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.468312 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-f84lw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.479889 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.484760 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.492188 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5z699" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.502183 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.503226 4718 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.529530 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.551024 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.564393 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.567151 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/64767824-fed2-496b-baa5-2c3ce0f8ee5b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.595140 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xknn\" (UniqueName: \"kubernetes.io/projected/96226f3a-94e6-4895-a9fc-662f07f8cdf8-kube-api-access-4xknn\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qsr\" (UID: \"96226f3a-94e6-4895-a9fc-662f07f8cdf8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.612717 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.615094 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87px2\" (UniqueName: \"kubernetes.io/projected/cbccb659-e964-4146-a586-d71b85f000f0-kube-api-access-87px2\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.626765 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2fhc\" (UniqueName: \"kubernetes.io/projected/8b619ef6-1db6-4a92-9810-aa00e60f5bb8-kube-api-access-r2fhc\") pod \"kube-storage-version-migrator-operator-b67b599dd-5kg75\" (UID: \"8b619ef6-1db6-4a92-9810-aa00e60f5bb8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.641911 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89k9s\" (UniqueName: \"kubernetes.io/projected/8204ea9d-c0c7-4541-8e4e-a1b96303aa54-kube-api-access-89k9s\") pod \"migrator-59844c95c7-c4lmg\" (UID: \"8204ea9d-c0c7-4541-8e4e-a1b96303aa54\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.658894 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfc4d\" (UniqueName: \"kubernetes.io/projected/daeca836-1e78-4156-9cbe-b065c3d2f59f-kube-api-access-qfc4d\") pod \"multus-admission-controller-857f4d67dd-vp8rk\" (UID: \"daeca836-1e78-4156-9cbe-b065c3d2f59f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.684877 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78pds\" (UniqueName: \"kubernetes.io/projected/64767824-fed2-496b-baa5-2c3ce0f8ee5b-kube-api-access-78pds\") pod \"cluster-image-registry-operator-dc59b4c8b-czh69\" (UID: \"64767824-fed2-496b-baa5-2c3ce0f8ee5b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.700995 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cbccb659-e964-4146-a586-d71b85f000f0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n76nm\" (UID: \"cbccb659-e964-4146-a586-d71b85f000f0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760303 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cb96422d-069d-48be-8a3d-da9a95a0e708-etcd-service-ca\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760708 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dz5r\" (UniqueName: \"kubernetes.io/projected/cb96422d-069d-48be-8a3d-da9a95a0e708-kube-api-access-6dz5r\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760740 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-service-ca-bundle\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760766 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/76e6b6a8-fa61-4a22-a1c6-54a244a13696-metrics-tls\") pod \"dns-default-7jwsr\" (UID: \"76e6b6a8-fa61-4a22-a1c6-54a244a13696\") " pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760808 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760825 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-default-certificate\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760859 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jzmpb\" (UID: \"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760875 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b47eec75-b43c-486d-be96-dce4faceddbf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760938 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-registry-certificates\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.760985 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b47eec75-b43c-486d-be96-dce4faceddbf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.761263 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29tws\" (UniqueName: \"kubernetes.io/projected/76e6b6a8-fa61-4a22-a1c6-54a244a13696-kube-api-access-29tws\") pod \"dns-default-7jwsr\" (UID: \"76e6b6a8-fa61-4a22-a1c6-54a244a13696\") " pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.761297 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3c8d3d9-76af-41a6-92b0-eee699956823-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rkfsp\" (UID: \"d3c8d3d9-76af-41a6-92b0-eee699956823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.761316 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb96422d-069d-48be-8a3d-da9a95a0e708-config\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.761354 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-bound-sa-token\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.761428 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76e6b6a8-fa61-4a22-a1c6-54a244a13696-config-volume\") pod \"dns-default-7jwsr\" (UID: \"76e6b6a8-fa61-4a22-a1c6-54a244a13696\") " pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:17 crc kubenswrapper[4718]: E1206 04:09:17.762146 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:18.262132351 +0000 UTC m=+147.267837512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.763887 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-metrics-certs\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.763919 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c8d3d9-76af-41a6-92b0-eee699956823-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rkfsp\" (UID: \"d3c8d3d9-76af-41a6-92b0-eee699956823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.768506 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmkqs\" (UniqueName: \"kubernetes.io/projected/1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82-kube-api-access-dmkqs\") pod \"machine-config-controller-84d6567774-jzmpb\" (UID: \"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.768855 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-registry-tls\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.768882 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb96422d-069d-48be-8a3d-da9a95a0e708-serving-cert\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.768918 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nknws\" (UniqueName: \"kubernetes.io/projected/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-kube-api-access-nknws\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.768938 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cb96422d-069d-48be-8a3d-da9a95a0e708-etcd-client\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.768955 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82-proxy-tls\") pod \"machine-config-controller-84d6567774-jzmpb\" (UID: \"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.768974 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c8d3d9-76af-41a6-92b0-eee699956823-config\") pod \"kube-apiserver-operator-766d6c64bb-rkfsp\" (UID: \"d3c8d3d9-76af-41a6-92b0-eee699956823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.769058 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-stats-auth\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.769099 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-trusted-ca\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.769115 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqxc2\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-kube-api-access-fqxc2\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.769146 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cb96422d-069d-48be-8a3d-da9a95a0e708-etcd-ca\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.846576 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.858918 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.872831 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873027 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6kwkv\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873049 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-apiservice-cert\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873087 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmkqs\" (UniqueName: \"kubernetes.io/projected/1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82-kube-api-access-dmkqs\") pod \"machine-config-controller-84d6567774-jzmpb\" (UID: \"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873113 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e3a6e268-54a8-47d6-823e-fc3476ac9364-proxy-tls\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873128 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk2f2\" (UniqueName: \"kubernetes.io/projected/0a1291c7-f265-4a3c-a994-3e8b3ab4e8df-kube-api-access-rk2f2\") pod \"catalog-operator-68c6474976-jn5bw\" (UID: \"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873143 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-plugins-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873161 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-registry-tls\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873178 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/982c800e-bebc-4214-a715-be8d91a05e83-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vnhhd\" (UID: \"982c800e-bebc-4214-a715-be8d91a05e83\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873211 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlh4s\" (UniqueName: \"kubernetes.io/projected/b4aca9cc-4401-4463-9560-95dae620f6eb-kube-api-access-qlh4s\") pod \"service-ca-operator-777779d784-vqswv\" (UID: \"b4aca9cc-4401-4463-9560-95dae620f6eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873261 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb96422d-069d-48be-8a3d-da9a95a0e708-serving-cert\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873278 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0a1291c7-f265-4a3c-a994-3e8b3ab4e8df-srv-cert\") pod \"catalog-operator-68c6474976-jn5bw\" (UID: \"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873293 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfcpf\" (UniqueName: \"kubernetes.io/projected/71558f8b-d394-4546-9c82-c9ef5f10a4c4-kube-api-access-zfcpf\") pod \"package-server-manager-789f6589d5-r5mv6\" (UID: \"71558f8b-d394-4546-9c82-c9ef5f10a4c4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873319 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nknws\" (UniqueName: \"kubernetes.io/projected/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-kube-api-access-nknws\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873334 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62glf\" (UniqueName: \"kubernetes.io/projected/d22fc175-392e-4fdb-86c3-5d78793275d1-kube-api-access-62glf\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873350 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85lss\" (UniqueName: \"kubernetes.io/projected/3b0daeef-042f-4c26-b228-f879326917b0-kube-api-access-85lss\") pod \"marketplace-operator-79b997595-6kwkv\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873366 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cb96422d-069d-48be-8a3d-da9a95a0e708-etcd-client\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873381 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73fa09b0-f084-4140-b014-febd78e2c2bf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h7222\" (UID: \"73fa09b0-f084-4140-b014-febd78e2c2bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873398 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82-proxy-tls\") pod \"machine-config-controller-84d6567774-jzmpb\" (UID: \"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873413 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fa09b0-f084-4140-b014-febd78e2c2bf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h7222\" (UID: \"73fa09b0-f084-4140-b014-febd78e2c2bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873429 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a287c864-444a-41c1-bcfc-d11ed8086aef-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-szgrx\" (UID: \"a287c864-444a-41c1-bcfc-d11ed8086aef\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873453 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-tmpfs\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873466 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-webhook-cert\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873501 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c8d3d9-76af-41a6-92b0-eee699956823-config\") pod \"kube-apiserver-operator-766d6c64bb-rkfsp\" (UID: \"d3c8d3d9-76af-41a6-92b0-eee699956823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873533 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/982c800e-bebc-4214-a715-be8d91a05e83-srv-cert\") pod \"olm-operator-6b444d44fb-vnhhd\" (UID: \"982c800e-bebc-4214-a715-be8d91a05e83\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873551 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cc87493a-5f71-450d-a4fa-0377185c077d-node-bootstrap-token\") pod \"machine-config-server-d96mp\" (UID: \"cc87493a-5f71-450d-a4fa-0377185c077d\") " pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873568 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-stats-auth\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873591 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt8sb\" (UniqueName: \"kubernetes.io/projected/d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf-kube-api-access-jt8sb\") pod \"ingress-canary-mxjzh\" (UID: \"d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf\") " pod="openshift-ingress-canary/ingress-canary-mxjzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873631 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7m9d\" (UniqueName: \"kubernetes.io/projected/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-kube-api-access-g7m9d\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873654 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-trusted-ca\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873670 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-secret-volume\") pod \"collect-profiles-29416560-c95jm\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873695 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqxc2\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-kube-api-access-fqxc2\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873709 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cb96422d-069d-48be-8a3d-da9a95a0e708-etcd-ca\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873723 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0a1291c7-f265-4a3c-a994-3e8b3ab4e8df-profile-collector-cert\") pod \"catalog-operator-68c6474976-jn5bw\" (UID: \"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873753 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-config-volume\") pod \"collect-profiles-29416560-c95jm\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873769 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lncxn\" (UniqueName: \"kubernetes.io/projected/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-kube-api-access-lncxn\") pod \"collect-profiles-29416560-c95jm\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873783 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnrdk\" (UniqueName: \"kubernetes.io/projected/ed46608d-0490-4b03-86ca-605ab4a57a93-kube-api-access-dnrdk\") pod \"service-ca-9c57cc56f-gfknw\" (UID: \"ed46608d-0490-4b03-86ca-605ab4a57a93\") " pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873797 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cc87493a-5f71-450d-a4fa-0377185c077d-certs\") pod \"machine-config-server-d96mp\" (UID: \"cc87493a-5f71-450d-a4fa-0377185c077d\") " pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873829 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2kcb\" (UniqueName: \"kubernetes.io/projected/cc87493a-5f71-450d-a4fa-0377185c077d-kube-api-access-q2kcb\") pod \"machine-config-server-d96mp\" (UID: \"cc87493a-5f71-450d-a4fa-0377185c077d\") " pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873844 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a287c864-444a-41c1-bcfc-d11ed8086aef-config\") pod \"kube-controller-manager-operator-78b949d7b-szgrx\" (UID: \"a287c864-444a-41c1-bcfc-d11ed8086aef\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873859 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vlt6\" (UniqueName: \"kubernetes.io/projected/982c800e-bebc-4214-a715-be8d91a05e83-kube-api-access-8vlt6\") pod \"olm-operator-6b444d44fb-vnhhd\" (UID: \"982c800e-bebc-4214-a715-be8d91a05e83\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873884 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4aca9cc-4401-4463-9560-95dae620f6eb-serving-cert\") pod \"service-ca-operator-777779d784-vqswv\" (UID: \"b4aca9cc-4401-4463-9560-95dae620f6eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873902 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cb96422d-069d-48be-8a3d-da9a95a0e708-etcd-service-ca\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873918 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e3a6e268-54a8-47d6-823e-fc3476ac9364-images\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873933 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4aca9cc-4401-4463-9560-95dae620f6eb-config\") pod \"service-ca-operator-777779d784-vqswv\" (UID: \"b4aca9cc-4401-4463-9560-95dae620f6eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873950 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dz5r\" (UniqueName: \"kubernetes.io/projected/cb96422d-069d-48be-8a3d-da9a95a0e708-kube-api-access-6dz5r\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873974 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-service-ca-bundle\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.873991 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/76e6b6a8-fa61-4a22-a1c6-54a244a13696-metrics-tls\") pod \"dns-default-7jwsr\" (UID: \"76e6b6a8-fa61-4a22-a1c6-54a244a13696\") " pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874006 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/71558f8b-d394-4546-9c82-c9ef5f10a4c4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r5mv6\" (UID: \"71558f8b-d394-4546-9c82-c9ef5f10a4c4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874030 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6kwkv\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874046 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-default-certificate\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874067 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-mountpoint-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874114 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jzmpb\" (UID: \"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874131 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73fa09b0-f084-4140-b014-febd78e2c2bf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h7222\" (UID: \"73fa09b0-f084-4140-b014-febd78e2c2bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874147 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b47eec75-b43c-486d-be96-dce4faceddbf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874180 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ed46608d-0490-4b03-86ca-605ab4a57a93-signing-key\") pod \"service-ca-9c57cc56f-gfknw\" (UID: \"ed46608d-0490-4b03-86ca-605ab4a57a93\") " pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874205 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-socket-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.874220 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf-cert\") pod \"ingress-canary-mxjzh\" (UID: \"d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf\") " pod="openshift-ingress-canary/ingress-canary-mxjzh" Dec 06 04:09:17 crc kubenswrapper[4718]: E1206 04:09:17.874302 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:18.374276193 +0000 UTC m=+147.379981354 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.879845 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cb96422d-069d-48be-8a3d-da9a95a0e708-etcd-ca\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.879852 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-service-ca-bundle\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.880492 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.881542 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-trusted-ca\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.881546 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c8d3d9-76af-41a6-92b0-eee699956823-config\") pod \"kube-apiserver-operator-766d6c64bb-rkfsp\" (UID: \"d3c8d3d9-76af-41a6-92b0-eee699956823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.881859 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b47eec75-b43c-486d-be96-dce4faceddbf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.882109 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jzmpb\" (UID: \"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.883031 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-registry-certificates\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.886593 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b47eec75-b43c-486d-be96-dce4faceddbf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.886727 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29tws\" (UniqueName: \"kubernetes.io/projected/76e6b6a8-fa61-4a22-a1c6-54a244a13696-kube-api-access-29tws\") pod \"dns-default-7jwsr\" (UID: \"76e6b6a8-fa61-4a22-a1c6-54a244a13696\") " pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.886803 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3c8d3d9-76af-41a6-92b0-eee699956823-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rkfsp\" (UID: \"d3c8d3d9-76af-41a6-92b0-eee699956823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.886889 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a287c864-444a-41c1-bcfc-d11ed8086aef-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-szgrx\" (UID: \"a287c864-444a-41c1-bcfc-d11ed8086aef\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.886958 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e3a6e268-54a8-47d6-823e-fc3476ac9364-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.887033 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ed46608d-0490-4b03-86ca-605ab4a57a93-signing-cabundle\") pod \"service-ca-9c57cc56f-gfknw\" (UID: \"ed46608d-0490-4b03-86ca-605ab4a57a93\") " pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.887111 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-csi-data-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.887181 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb96422d-069d-48be-8a3d-da9a95a0e708-config\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.887277 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-bound-sa-token\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.887352 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rf4r\" (UniqueName: \"kubernetes.io/projected/e3a6e268-54a8-47d6-823e-fc3476ac9364-kube-api-access-9rf4r\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.887451 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76e6b6a8-fa61-4a22-a1c6-54a244a13696-config-volume\") pod \"dns-default-7jwsr\" (UID: \"76e6b6a8-fa61-4a22-a1c6-54a244a13696\") " pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.887575 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-metrics-certs\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.887647 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c8d3d9-76af-41a6-92b0-eee699956823-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rkfsp\" (UID: \"d3c8d3d9-76af-41a6-92b0-eee699956823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.887717 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-registration-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.884017 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-registry-certificates\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.889796 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb96422d-069d-48be-8a3d-da9a95a0e708-config\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.884769 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cb96422d-069d-48be-8a3d-da9a95a0e708-etcd-service-ca\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.890792 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76e6b6a8-fa61-4a22-a1c6-54a244a13696-config-volume\") pod \"dns-default-7jwsr\" (UID: \"76e6b6a8-fa61-4a22-a1c6-54a244a13696\") " pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.892462 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82-proxy-tls\") pod \"machine-config-controller-84d6567774-jzmpb\" (UID: \"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.896730 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-default-certificate\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.898866 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b47eec75-b43c-486d-be96-dce4faceddbf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.902958 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-registry-tls\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.903289 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.903749 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cb96422d-069d-48be-8a3d-da9a95a0e708-etcd-client\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.912034 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-metrics-certs\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.916143 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-stats-auth\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.916647 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/76e6b6a8-fa61-4a22-a1c6-54a244a13696-metrics-tls\") pod \"dns-default-7jwsr\" (UID: \"76e6b6a8-fa61-4a22-a1c6-54a244a13696\") " pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.920567 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb96422d-069d-48be-8a3d-da9a95a0e708-serving-cert\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.927668 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c8d3d9-76af-41a6-92b0-eee699956823-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rkfsp\" (UID: \"d3c8d3d9-76af-41a6-92b0-eee699956823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.930845 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmkqs\" (UniqueName: \"kubernetes.io/projected/1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82-kube-api-access-dmkqs\") pod \"machine-config-controller-84d6567774-jzmpb\" (UID: \"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.940438 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.943201 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqxc2\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-kube-api-access-fqxc2\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.966632 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb297"] Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.966974 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.978996 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nknws\" (UniqueName: \"kubernetes.io/projected/8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5-kube-api-access-nknws\") pod \"router-default-5444994796-w4fdl\" (UID: \"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5\") " pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.988872 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e3a6e268-54a8-47d6-823e-fc3476ac9364-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.988923 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a287c864-444a-41c1-bcfc-d11ed8086aef-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-szgrx\" (UID: \"a287c864-444a-41c1-bcfc-d11ed8086aef\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.988941 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-csi-data-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.988956 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ed46608d-0490-4b03-86ca-605ab4a57a93-signing-cabundle\") pod \"service-ca-9c57cc56f-gfknw\" (UID: \"ed46608d-0490-4b03-86ca-605ab4a57a93\") " pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.988977 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rf4r\" (UniqueName: \"kubernetes.io/projected/e3a6e268-54a8-47d6-823e-fc3476ac9364-kube-api-access-9rf4r\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.988997 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-registration-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989021 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6kwkv\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989037 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-apiservice-cert\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989054 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e3a6e268-54a8-47d6-823e-fc3476ac9364-proxy-tls\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989071 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk2f2\" (UniqueName: \"kubernetes.io/projected/0a1291c7-f265-4a3c-a994-3e8b3ab4e8df-kube-api-access-rk2f2\") pod \"catalog-operator-68c6474976-jn5bw\" (UID: \"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989088 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-plugins-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989104 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/982c800e-bebc-4214-a715-be8d91a05e83-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vnhhd\" (UID: \"982c800e-bebc-4214-a715-be8d91a05e83\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989121 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlh4s\" (UniqueName: \"kubernetes.io/projected/b4aca9cc-4401-4463-9560-95dae620f6eb-kube-api-access-qlh4s\") pod \"service-ca-operator-777779d784-vqswv\" (UID: \"b4aca9cc-4401-4463-9560-95dae620f6eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989135 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0a1291c7-f265-4a3c-a994-3e8b3ab4e8df-srv-cert\") pod \"catalog-operator-68c6474976-jn5bw\" (UID: \"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989150 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfcpf\" (UniqueName: \"kubernetes.io/projected/71558f8b-d394-4546-9c82-c9ef5f10a4c4-kube-api-access-zfcpf\") pod \"package-server-manager-789f6589d5-r5mv6\" (UID: \"71558f8b-d394-4546-9c82-c9ef5f10a4c4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989165 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62glf\" (UniqueName: \"kubernetes.io/projected/d22fc175-392e-4fdb-86c3-5d78793275d1-kube-api-access-62glf\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989179 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85lss\" (UniqueName: \"kubernetes.io/projected/3b0daeef-042f-4c26-b228-f879326917b0-kube-api-access-85lss\") pod \"marketplace-operator-79b997595-6kwkv\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989197 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73fa09b0-f084-4140-b014-febd78e2c2bf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h7222\" (UID: \"73fa09b0-f084-4140-b014-febd78e2c2bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989211 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fa09b0-f084-4140-b014-febd78e2c2bf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h7222\" (UID: \"73fa09b0-f084-4140-b014-febd78e2c2bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989224 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a287c864-444a-41c1-bcfc-d11ed8086aef-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-szgrx\" (UID: \"a287c864-444a-41c1-bcfc-d11ed8086aef\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989254 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-tmpfs\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989269 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-webhook-cert\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989295 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/982c800e-bebc-4214-a715-be8d91a05e83-srv-cert\") pod \"olm-operator-6b444d44fb-vnhhd\" (UID: \"982c800e-bebc-4214-a715-be8d91a05e83\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989315 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cc87493a-5f71-450d-a4fa-0377185c077d-node-bootstrap-token\") pod \"machine-config-server-d96mp\" (UID: \"cc87493a-5f71-450d-a4fa-0377185c077d\") " pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989334 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt8sb\" (UniqueName: \"kubernetes.io/projected/d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf-kube-api-access-jt8sb\") pod \"ingress-canary-mxjzh\" (UID: \"d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf\") " pod="openshift-ingress-canary/ingress-canary-mxjzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989350 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7m9d\" (UniqueName: \"kubernetes.io/projected/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-kube-api-access-g7m9d\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989365 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-secret-volume\") pod \"collect-profiles-29416560-c95jm\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989380 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0a1291c7-f265-4a3c-a994-3e8b3ab4e8df-profile-collector-cert\") pod \"catalog-operator-68c6474976-jn5bw\" (UID: \"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989395 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-config-volume\") pod \"collect-profiles-29416560-c95jm\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989411 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lncxn\" (UniqueName: \"kubernetes.io/projected/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-kube-api-access-lncxn\") pod \"collect-profiles-29416560-c95jm\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989426 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnrdk\" (UniqueName: \"kubernetes.io/projected/ed46608d-0490-4b03-86ca-605ab4a57a93-kube-api-access-dnrdk\") pod \"service-ca-9c57cc56f-gfknw\" (UID: \"ed46608d-0490-4b03-86ca-605ab4a57a93\") " pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989440 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cc87493a-5f71-450d-a4fa-0377185c077d-certs\") pod \"machine-config-server-d96mp\" (UID: \"cc87493a-5f71-450d-a4fa-0377185c077d\") " pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989454 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2kcb\" (UniqueName: \"kubernetes.io/projected/cc87493a-5f71-450d-a4fa-0377185c077d-kube-api-access-q2kcb\") pod \"machine-config-server-d96mp\" (UID: \"cc87493a-5f71-450d-a4fa-0377185c077d\") " pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989472 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a287c864-444a-41c1-bcfc-d11ed8086aef-config\") pod \"kube-controller-manager-operator-78b949d7b-szgrx\" (UID: \"a287c864-444a-41c1-bcfc-d11ed8086aef\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989488 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vlt6\" (UniqueName: \"kubernetes.io/projected/982c800e-bebc-4214-a715-be8d91a05e83-kube-api-access-8vlt6\") pod \"olm-operator-6b444d44fb-vnhhd\" (UID: \"982c800e-bebc-4214-a715-be8d91a05e83\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989506 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4aca9cc-4401-4463-9560-95dae620f6eb-serving-cert\") pod \"service-ca-operator-777779d784-vqswv\" (UID: \"b4aca9cc-4401-4463-9560-95dae620f6eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989521 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4aca9cc-4401-4463-9560-95dae620f6eb-config\") pod \"service-ca-operator-777779d784-vqswv\" (UID: \"b4aca9cc-4401-4463-9560-95dae620f6eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989538 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e3a6e268-54a8-47d6-823e-fc3476ac9364-images\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989562 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/71558f8b-d394-4546-9c82-c9ef5f10a4c4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r5mv6\" (UID: \"71558f8b-d394-4546-9c82-c9ef5f10a4c4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989582 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6kwkv\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989601 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989617 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-mountpoint-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989637 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73fa09b0-f084-4140-b014-febd78e2c2bf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h7222\" (UID: \"73fa09b0-f084-4140-b014-febd78e2c2bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989652 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ed46608d-0490-4b03-86ca-605ab4a57a93-signing-key\") pod \"service-ca-9c57cc56f-gfknw\" (UID: \"ed46608d-0490-4b03-86ca-605ab4a57a93\") " pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989667 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-socket-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.989684 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf-cert\") pod \"ingress-canary-mxjzh\" (UID: \"d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf\") " pod="openshift-ingress-canary/ingress-canary-mxjzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.994219 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf-cert\") pod \"ingress-canary-mxjzh\" (UID: \"d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf\") " pod="openshift-ingress-canary/ingress-canary-mxjzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.995346 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-plugins-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.995626 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dz5r\" (UniqueName: \"kubernetes.io/projected/cb96422d-069d-48be-8a3d-da9a95a0e708-kube-api-access-6dz5r\") pod \"etcd-operator-b45778765-x2dzh\" (UID: \"cb96422d-069d-48be-8a3d-da9a95a0e708\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.995906 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e3a6e268-54a8-47d6-823e-fc3476ac9364-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.996356 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-config-volume\") pod \"collect-profiles-29416560-c95jm\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.996772 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a287c864-444a-41c1-bcfc-d11ed8086aef-config\") pod \"kube-controller-manager-operator-78b949d7b-szgrx\" (UID: \"a287c864-444a-41c1-bcfc-d11ed8086aef\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.996841 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-registration-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.996941 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-csi-data-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.997622 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ed46608d-0490-4b03-86ca-605ab4a57a93-signing-cabundle\") pod \"service-ca-9c57cc56f-gfknw\" (UID: \"ed46608d-0490-4b03-86ca-605ab4a57a93\") " pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.997929 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/982c800e-bebc-4214-a715-be8d91a05e83-srv-cert\") pod \"olm-operator-6b444d44fb-vnhhd\" (UID: \"982c800e-bebc-4214-a715-be8d91a05e83\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.998282 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-tmpfs\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.999066 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fa09b0-f084-4140-b014-febd78e2c2bf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h7222\" (UID: \"73fa09b0-f084-4140-b014-febd78e2c2bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.999226 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4aca9cc-4401-4463-9560-95dae620f6eb-config\") pod \"service-ca-operator-777779d784-vqswv\" (UID: \"b4aca9cc-4401-4463-9560-95dae620f6eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:17 crc kubenswrapper[4718]: I1206 04:09:17.999357 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e3a6e268-54a8-47d6-823e-fc3476ac9364-images\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:17 crc kubenswrapper[4718]: E1206 04:09:17.999430 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:18.499418493 +0000 UTC m=+147.505123654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.000034 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-socket-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.000163 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d22fc175-392e-4fdb-86c3-5d78793275d1-mountpoint-dir\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.000893 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29tws\" (UniqueName: \"kubernetes.io/projected/76e6b6a8-fa61-4a22-a1c6-54a244a13696-kube-api-access-29tws\") pod \"dns-default-7jwsr\" (UID: \"76e6b6a8-fa61-4a22-a1c6-54a244a13696\") " pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.001349 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6kwkv\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.002783 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0a1291c7-f265-4a3c-a994-3e8b3ab4e8df-profile-collector-cert\") pod \"catalog-operator-68c6474976-jn5bw\" (UID: \"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.004454 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-webhook-cert\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.005483 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-apiservice-cert\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.008725 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-secret-volume\") pod \"collect-profiles-29416560-c95jm\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.009203 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4aca9cc-4401-4463-9560-95dae620f6eb-serving-cert\") pod \"service-ca-operator-777779d784-vqswv\" (UID: \"b4aca9cc-4401-4463-9560-95dae620f6eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.009870 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/71558f8b-d394-4546-9c82-c9ef5f10a4c4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r5mv6\" (UID: \"71558f8b-d394-4546-9c82-c9ef5f10a4c4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.012034 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cc87493a-5f71-450d-a4fa-0377185c077d-node-bootstrap-token\") pod \"machine-config-server-d96mp\" (UID: \"cc87493a-5f71-450d-a4fa-0377185c077d\") " pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.012205 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cc87493a-5f71-450d-a4fa-0377185c077d-certs\") pod \"machine-config-server-d96mp\" (UID: \"cc87493a-5f71-450d-a4fa-0377185c077d\") " pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.014630 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73fa09b0-f084-4140-b014-febd78e2c2bf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h7222\" (UID: \"73fa09b0-f084-4140-b014-febd78e2c2bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.018419 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6kwkv\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.021199 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3c8d3d9-76af-41a6-92b0-eee699956823-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rkfsp\" (UID: \"d3c8d3d9-76af-41a6-92b0-eee699956823\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.021724 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/982c800e-bebc-4214-a715-be8d91a05e83-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vnhhd\" (UID: \"982c800e-bebc-4214-a715-be8d91a05e83\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.021832 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0a1291c7-f265-4a3c-a994-3e8b3ab4e8df-srv-cert\") pod \"catalog-operator-68c6474976-jn5bw\" (UID: \"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.022196 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e3a6e268-54a8-47d6-823e-fc3476ac9364-proxy-tls\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.026531 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a287c864-444a-41c1-bcfc-d11ed8086aef-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-szgrx\" (UID: \"a287c864-444a-41c1-bcfc-d11ed8086aef\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.027793 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ed46608d-0490-4b03-86ca-605ab4a57a93-signing-key\") pod \"service-ca-9c57cc56f-gfknw\" (UID: \"ed46608d-0490-4b03-86ca-605ab4a57a93\") " pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.034732 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-bound-sa-token\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.086558 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vlt6\" (UniqueName: \"kubernetes.io/projected/982c800e-bebc-4214-a715-be8d91a05e83-kube-api-access-8vlt6\") pod \"olm-operator-6b444d44fb-vnhhd\" (UID: \"982c800e-bebc-4214-a715-be8d91a05e83\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.093253 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.093927 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:18.593910587 +0000 UTC m=+147.599615748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.105741 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.106572 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lncxn\" (UniqueName: \"kubernetes.io/projected/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-kube-api-access-lncxn\") pod \"collect-profiles-29416560-c95jm\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.123643 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" event={"ID":"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a","Type":"ContainerStarted","Data":"95eaac06549b409c9d323ab1f97442e4314a6c77799b463c9feeefdff41bb7b5"} Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.138976 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnrdk\" (UniqueName: \"kubernetes.io/projected/ed46608d-0490-4b03-86ca-605ab4a57a93-kube-api-access-dnrdk\") pod \"service-ca-9c57cc56f-gfknw\" (UID: \"ed46608d-0490-4b03-86ca-605ab4a57a93\") " pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.158501 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" event={"ID":"79292fb4-57be-4bbe-aa46-09d97f06b109","Type":"ContainerStarted","Data":"db55c972e5ca8c74e309d2ed8c82100c91767d71510807c82f08e70807c597b7"} Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.162260 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85lss\" (UniqueName: \"kubernetes.io/projected/3b0daeef-042f-4c26-b228-f879326917b0-kube-api-access-85lss\") pod \"marketplace-operator-79b997595-6kwkv\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.175804 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2kcb\" (UniqueName: \"kubernetes.io/projected/cc87493a-5f71-450d-a4fa-0377185c077d-kube-api-access-q2kcb\") pod \"machine-config-server-d96mp\" (UID: \"cc87493a-5f71-450d-a4fa-0377185c077d\") " pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.181567 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" event={"ID":"7412230b-433d-4a8b-86d7-38cad91e6ecb","Type":"ContainerStarted","Data":"18fc4c18230dd05d7c92bc5882314e66a08fbab22461994a7a74db83619201ab"} Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.181609 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" event={"ID":"7412230b-433d-4a8b-86d7-38cad91e6ecb","Type":"ContainerStarted","Data":"9972c0e3ba1c83ec2cd64eca1dda87c40f365025d36fc81259716def803b51aa"} Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.185988 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1e2ee2c24ad5a38f78d929bc4752f63a5f0c1beef4eaa10a5af5cfe0df00033e"} Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.204318 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.207472 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a287c864-444a-41c1-bcfc-d11ed8086aef-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-szgrx\" (UID: \"a287c864-444a-41c1-bcfc-d11ed8086aef\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.208897 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:18.70888067 +0000 UTC m=+147.714585831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.213588 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.220740 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.225892 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt8sb\" (UniqueName: \"kubernetes.io/projected/d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf-kube-api-access-jt8sb\") pod \"ingress-canary-mxjzh\" (UID: \"d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf\") " pod="openshift-ingress-canary/ingress-canary-mxjzh" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.236276 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.247544 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7m9d\" (UniqueName: \"kubernetes.io/projected/7cf9c034-70ae-4c39-aca8-cc0a71a7a850-kube-api-access-g7m9d\") pod \"packageserver-d55dfcdfc-r4jb7\" (UID: \"7cf9c034-70ae-4c39-aca8-cc0a71a7a850\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.261654 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rf4r\" (UniqueName: \"kubernetes.io/projected/e3a6e268-54a8-47d6-823e-fc3476ac9364-kube-api-access-9rf4r\") pod \"machine-config-operator-74547568cd-z7fbp\" (UID: \"e3a6e268-54a8-47d6-823e-fc3476ac9364\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.268369 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlh4s\" (UniqueName: \"kubernetes.io/projected/b4aca9cc-4401-4463-9560-95dae620f6eb-kube-api-access-qlh4s\") pod \"service-ca-operator-777779d784-vqswv\" (UID: \"b4aca9cc-4401-4463-9560-95dae620f6eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.272924 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.294022 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2"] Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.305859 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.306873 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.307415 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73fa09b0-f084-4140-b014-febd78e2c2bf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h7222\" (UID: \"73fa09b0-f084-4140-b014-febd78e2c2bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.313819 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.314013 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:18.813991846 +0000 UTC m=+147.819697007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.314218 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.314714 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-d96mp" Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.315966 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:18.81595128 +0000 UTC m=+147.821656441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.325183 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.332086 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.332635 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.340840 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.354764 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mxjzh" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.355220 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.360494 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfcpf\" (UniqueName: \"kubernetes.io/projected/71558f8b-d394-4546-9c82-c9ef5f10a4c4-kube-api-access-zfcpf\") pod \"package-server-manager-789f6589d5-r5mv6\" (UID: \"71558f8b-d394-4546-9c82-c9ef5f10a4c4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.364961 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62glf\" (UniqueName: \"kubernetes.io/projected/d22fc175-392e-4fdb-86c3-5d78793275d1-kube-api-access-62glf\") pod \"csi-hostpathplugin-59sd7\" (UID: \"d22fc175-392e-4fdb-86c3-5d78793275d1\") " pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.367015 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk2f2\" (UniqueName: \"kubernetes.io/projected/0a1291c7-f265-4a3c-a994-3e8b3ab4e8df-kube-api-access-rk2f2\") pod \"catalog-operator-68c6474976-jn5bw\" (UID: \"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.394208 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-59sd7" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.417529 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.417757 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:18.917740889 +0000 UTC m=+147.923446050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.417807 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.418093 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:18.91808479 +0000 UTC m=+147.923789951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.450983 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2zd9s"] Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.484666 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.494301 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.518601 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.519060 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.019045593 +0000 UTC m=+148.024750754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.583568 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.590597 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.621613 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.621943 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.121932687 +0000 UTC m=+148.127637848 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.722941 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.723312 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.223293884 +0000 UTC m=+148.228999045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.826010 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.826416 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.326401065 +0000 UTC m=+148.332106216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.833769 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xmfbd" podStartSLOduration=128.833750292 podStartE2EDuration="2m8.833750292s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:18.83276316 +0000 UTC m=+147.838468321" watchObservedRunningTime="2025-12-06 04:09:18.833750292 +0000 UTC m=+147.839455453" Dec 06 04:09:18 crc kubenswrapper[4718]: I1206 04:09:18.928919 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:18 crc kubenswrapper[4718]: E1206 04:09:18.929268 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.429253579 +0000 UTC m=+148.434958740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.012996 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-f84lw"] Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.032068 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.032516 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.532502376 +0000 UTC m=+148.538207537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.055432 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-29cng"] Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.057371 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws"] Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.076758 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx"] Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.079622 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8"] Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.134801 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.135442 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.635422401 +0000 UTC m=+148.641127562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.231482 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" event={"ID":"77330fe5-fc23-41b0-9b9d-e6497908bd77","Type":"ContainerStarted","Data":"2757cceb08d6705a70b0c0499af6c9a0147818ee1ccda042533051cc222ae843"} Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.241937 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.242380 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.742365286 +0000 UTC m=+148.748070447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.273737 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-d96mp" event={"ID":"cc87493a-5f71-450d-a4fa-0377185c077d","Type":"ContainerStarted","Data":"3c53a5d5cceae5b38d962d8ab8821a94ab40643fe8df08a6174897fd0b33068b"} Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.299809 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w4fdl" event={"ID":"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5","Type":"ContainerStarted","Data":"518083eb41b6026c6e67a897131773a50c9dc60ae48b96fe897c96dd6b1652c3"} Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.312463 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" event={"ID":"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe","Type":"ContainerStarted","Data":"3a11467dc8839fcb7c40ed7d177330a971df059f93db81e30eca02871e5ac7dc"} Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.344850 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.345965 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.845949483 +0000 UTC m=+148.851654644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.354965 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"cc10c5445129a730c045d11109134ed7b2bb570cd915d444f604d188c5a9f402"} Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.451288 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.453070 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:19.953056024 +0000 UTC m=+148.958761185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.553802 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.554104 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.054089939 +0000 UTC m=+149.059795100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.574168 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" podStartSLOduration=129.574142192 podStartE2EDuration="2m9.574142192s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:19.528741234 +0000 UTC m=+148.534446395" watchObservedRunningTime="2025-12-06 04:09:19.574142192 +0000 UTC m=+148.579847353" Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.607459 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-6jmn4" podStartSLOduration=129.607432202 podStartE2EDuration="2m9.607432202s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:19.559151332 +0000 UTC m=+148.564856503" watchObservedRunningTime="2025-12-06 04:09:19.607432202 +0000 UTC m=+148.613137363" Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.655109 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.655641 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.15561003 +0000 UTC m=+149.161315191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.702573 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" podStartSLOduration=129.702544817 podStartE2EDuration="2m9.702544817s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:19.69826564 +0000 UTC m=+148.703970801" watchObservedRunningTime="2025-12-06 04:09:19.702544817 +0000 UTC m=+148.708249978" Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.756372 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.756476 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.256454109 +0000 UTC m=+149.262159270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.756726 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.757051 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.257038227 +0000 UTC m=+149.262743388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.860063 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.865435 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.365404168 +0000 UTC m=+149.371109329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.869872 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.884305 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.384286985 +0000 UTC m=+149.389992146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.893419 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-69v98"] Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.935510 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" podStartSLOduration=129.935481169 podStartE2EDuration="2m9.935481169s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:19.929261849 +0000 UTC m=+148.934967010" watchObservedRunningTime="2025-12-06 04:09:19.935481169 +0000 UTC m=+148.941186330" Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.970886 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:19 crc kubenswrapper[4718]: E1206 04:09:19.971200 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.471185676 +0000 UTC m=+149.476890837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:19 crc kubenswrapper[4718]: I1206 04:09:19.995318 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" podStartSLOduration=129.99529803 podStartE2EDuration="2m9.99529803s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:19.989896716 +0000 UTC m=+148.995601877" watchObservedRunningTime="2025-12-06 04:09:19.99529803 +0000 UTC m=+149.001003201" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.024355 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5z699"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.038155 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-w4fdl" podStartSLOduration=130.038135976 podStartE2EDuration="2m10.038135976s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.03329345 +0000 UTC m=+149.038998631" watchObservedRunningTime="2025-12-06 04:09:20.038135976 +0000 UTC m=+149.043841137" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.065137 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-v6j6g"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.072667 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.073131 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.57311549 +0000 UTC m=+149.578820651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.177173 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.177527 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.677509793 +0000 UTC m=+149.683214954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.197027 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69"] Dec 06 04:09:20 crc kubenswrapper[4718]: W1206 04:09:20.214511 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64767824_fed2_496b_baa5_2c3ce0f8ee5b.slice/crio-14ddf74b2ec2223d906974b4fac4f03994f184073f4487aa778d18956c555cea WatchSource:0}: Error finding container 14ddf74b2ec2223d906974b4fac4f03994f184073f4487aa778d18956c555cea: Status 404 returned error can't find the container with id 14ddf74b2ec2223d906974b4fac4f03994f184073f4487aa778d18956c555cea Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.224318 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.232560 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.234133 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:20 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:20 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:20 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.234214 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.240972 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-x2dzh"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.275664 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.279430 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.279748 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.779733236 +0000 UTC m=+149.785438397 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.365197 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-d96mp" event={"ID":"cc87493a-5f71-450d-a4fa-0377185c077d","Type":"ContainerStarted","Data":"31091bc7c49c6f7c8f760c04387515f975dbd65424bb233115dd15726f66accc"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.368123 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-69v98" event={"ID":"f4ffee8e-d1ea-4908-b160-473b1a73d31c","Type":"ContainerStarted","Data":"df57ad08c9e444501ef26c2efce8fbfb237da0566b3f63ae66c931762e357d65"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.368152 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-69v98" event={"ID":"f4ffee8e-d1ea-4908-b160-473b1a73d31c","Type":"ContainerStarted","Data":"bc8918538422c5a90a5b2f9e3160d64cb8ca97c19f6fbae9bb09f41ca3e8ee89"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.369024 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.371487 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" event={"ID":"cb96422d-069d-48be-8a3d-da9a95a0e708","Type":"ContainerStarted","Data":"5e02e21348f872684ebf04672b5e60ddc26248e643f4a76b16569c64b99b2d72"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.374378 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" event={"ID":"77330fe5-fc23-41b0-9b9d-e6497908bd77","Type":"ContainerStarted","Data":"1c0bc164b26d3ca7fa048799ae214f903a246ff1a255f0fb189ab29c06d2af57"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.377541 4718 patch_prober.go:28] interesting pod/console-operator-58897d9998-69v98 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.377590 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-69v98" podUID="f4ffee8e-d1ea-4908-b160-473b1a73d31c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.382670 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-d96mp" podStartSLOduration=5.382621121 podStartE2EDuration="5.382621121s" podCreationTimestamp="2025-12-06 04:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.379885403 +0000 UTC m=+149.385590564" watchObservedRunningTime="2025-12-06 04:09:20.382621121 +0000 UTC m=+149.388326282" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.385318 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.385872 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.885843975 +0000 UTC m=+149.891549146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.386271 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.394141 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.894043538 +0000 UTC m=+149.899748699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.410588 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"16389efd87457af132415548dd5141560c8daf96a0b4cd768dfc47d16bebb6c7"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.421577 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" event={"ID":"64767824-fed2-496b-baa5-2c3ce0f8ee5b","Type":"ContainerStarted","Data":"14ddf74b2ec2223d906974b4fac4f03994f184073f4487aa778d18956c555cea"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.448638 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-69v98" podStartSLOduration=130.448617361 podStartE2EDuration="2m10.448617361s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.448415805 +0000 UTC m=+149.454120996" watchObservedRunningTime="2025-12-06 04:09:20.448617361 +0000 UTC m=+149.454322532" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.456950 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.466075 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.466940 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5z699" event={"ID":"e812f183-3419-4365-99f5-94f1e9f8abcb","Type":"ContainerStarted","Data":"c1b3acb3d7c757cc8b94ebb722e604015acb23bbdffb106f393db7ed593a14c4"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.481195 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2zd9s" podStartSLOduration=130.481172867 podStartE2EDuration="2m10.481172867s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.478049756 +0000 UTC m=+149.483754917" watchObservedRunningTime="2025-12-06 04:09:20.481172867 +0000 UTC m=+149.486878028" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.482407 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vp8rk"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.482453 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-f84lw" event={"ID":"1308cc0b-85c6-49e9-a036-fc110f461f9c","Type":"ContainerStarted","Data":"ff75bf8a86aed0d540948640a4e4d1325327cbf05d834b8f1ad2f6a622072a31"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.482487 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-f84lw" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.482500 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-f84lw" event={"ID":"1308cc0b-85c6-49e9-a036-fc110f461f9c","Type":"ContainerStarted","Data":"548d2b3170f18ae02e104b258120f22aa0a0071b26acb9c8c0962369a829ed52"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.494840 4718 patch_prober.go:28] interesting pod/downloads-7954f5f757-f84lw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.495106 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-f84lw" podUID="1308cc0b-85c6-49e9-a036-fc110f461f9c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.495393 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.495893 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.995875669 +0000 UTC m=+150.001580830 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.496206 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.498141 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222"] Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.499194 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:20.999173575 +0000 UTC m=+150.004878726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.501079 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" event={"ID":"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5","Type":"ContainerStarted","Data":"63b2825f010ab64c357f67d18de679ff22fd37b234a3d5ef976352032ed0e1ca"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.502430 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" event={"ID":"ad88155c-cf04-4c1b-b92f-3c0bbd0b61e5","Type":"ContainerStarted","Data":"21f2a9e003e5758e3d3e6913652b4866af06029d35423170e2624b20ab60c11f"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.504462 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-74rb2" event={"ID":"83164f34-cd50-4f7d-b6ae-f9e65d3b69fe","Type":"ContainerStarted","Data":"d884052d7de8b8ec915a7d9125aaaa56f3047768347fec1c06976e76afdc0ba9"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.504778 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-59sd7"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.528061 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" event={"ID":"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a","Type":"ContainerStarted","Data":"0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.529499 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.540503 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.540536 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.542564 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-f84lw" podStartSLOduration=130.542540448 podStartE2EDuration="2m10.542540448s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.527650519 +0000 UTC m=+149.533355680" watchObservedRunningTime="2025-12-06 04:09:20.542540448 +0000 UTC m=+149.548245629" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.554809 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gfknw"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.555555 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" event={"ID":"96226f3a-94e6-4895-a9fc-662f07f8cdf8","Type":"ContainerStarted","Data":"062a006e42d6ca8eff58e5e5351e29eaaf95a376e0d253947664d211dcb412e1"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.557613 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vqswv"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.577365 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" podStartSLOduration=130.577340126 podStartE2EDuration="2m10.577340126s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.561539639 +0000 UTC m=+149.567244810" watchObservedRunningTime="2025-12-06 04:09:20.577340126 +0000 UTC m=+149.583045287" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.587093 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w4fdl" event={"ID":"8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5","Type":"ContainerStarted","Data":"f6e3691c4f9b91400c7f416be12620f464b1e05ffbb29b6d369bbce631a9f9a9"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.589384 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lzvl8" podStartSLOduration=130.589366431 podStartE2EDuration="2m10.589366431s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.586375566 +0000 UTC m=+149.592080737" watchObservedRunningTime="2025-12-06 04:09:20.589366431 +0000 UTC m=+149.595071592" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.598420 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.598643 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v6j6g" event={"ID":"454877ff-cab8-4340-bfe7-1709bbd1227d","Type":"ContainerStarted","Data":"b74f022b5eed208fdcde80800f8a4d5f34b6b5951420007d9da124ac965e36d3"} Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.600145 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.100127817 +0000 UTC m=+150.105832978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.602956 4718 generic.go:334] "Generic (PLEG): container finished" podID="58ce50cf-28a9-43a5-b4b3-bee39bbf7393" containerID="664ccf52b86ba945c9dcdc0b78d16fa27846b565fb9be19bc387c734eb7a6e82" exitCode=0 Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.603039 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" event={"ID":"58ce50cf-28a9-43a5-b4b3-bee39bbf7393","Type":"ContainerDied","Data":"664ccf52b86ba945c9dcdc0b78d16fa27846b565fb9be19bc387c734eb7a6e82"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.603073 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" event={"ID":"58ce50cf-28a9-43a5-b4b3-bee39bbf7393","Type":"ContainerStarted","Data":"ad473179ef4ed0e8ba9835e2eac703dbd698f9fb701e71c36e9af4072188eda0"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.606365 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" event={"ID":"cbccb659-e964-4146-a586-d71b85f000f0","Type":"ContainerStarted","Data":"c7ad40c5cd9918678638bda3df7c0f97ab959dd2ccd685226316b009e269a6cf"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.631131 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" event={"ID":"7ef554fa-0472-493f-b4ac-c131e53eb901","Type":"ContainerStarted","Data":"633317b7f9a3c90ab53c8a4590bd7f31a5f89d1739e5af3179a63198b4370c8d"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.631174 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" event={"ID":"7ef554fa-0472-493f-b4ac-c131e53eb901","Type":"ContainerStarted","Data":"6cfad270acf055d32cb10ad26ef841ec52292ed03f430145c3fccf20e123175e"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.631184 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" event={"ID":"7ef554fa-0472-493f-b4ac-c131e53eb901","Type":"ContainerStarted","Data":"2eebf04e82d1be0b90d2a25df498bc09bbdeafb730797fb1091353312e5be063"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.640519 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" event={"ID":"c76e70f2-707c-4d55-8af2-672fac1f7ac4","Type":"ContainerStarted","Data":"6ffed2f61374089653bb845fb2e38803382615ca3ea948ed4b24a4b221627811"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.640565 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.640575 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" event={"ID":"c76e70f2-707c-4d55-8af2-672fac1f7ac4","Type":"ContainerStarted","Data":"1696c6b496fd16d0f3365033ce45862d2f40359388f21ded797563687cb94687"} Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.648336 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6kwkv"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.670699 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mxjzh"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.676879 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-v6j6g" podStartSLOduration=130.676850492 podStartE2EDuration="2m10.676850492s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.675146937 +0000 UTC m=+149.680852108" watchObservedRunningTime="2025-12-06 04:09:20.676850492 +0000 UTC m=+149.682555673" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.685861 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7jwsr"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.702611 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.704517 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.20449366 +0000 UTC m=+150.210198821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.713350 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.718296 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.721973 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp"] Dec 06 04:09:20 crc kubenswrapper[4718]: W1206 04:09:20.723199 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded46608d_0490_4b03_86ca_605ab4a57a93.slice/crio-5d7ff2b158058445eb3d264762aa6c047f6f546077341218e83b1b213ecdc506 WatchSource:0}: Error finding container 5d7ff2b158058445eb3d264762aa6c047f6f546077341218e83b1b213ecdc506: Status 404 returned error can't find the container with id 5d7ff2b158058445eb3d264762aa6c047f6f546077341218e83b1b213ecdc506 Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.727315 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" podStartSLOduration=129.727294312 podStartE2EDuration="2m9.727294312s" podCreationTimestamp="2025-12-06 04:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.70793186 +0000 UTC m=+149.713637031" watchObservedRunningTime="2025-12-06 04:09:20.727294312 +0000 UTC m=+149.732999473" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.734399 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.751096 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-kwhws" podStartSLOduration=130.751074316 podStartE2EDuration="2m10.751074316s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:20.746542811 +0000 UTC m=+149.752247972" watchObservedRunningTime="2025-12-06 04:09:20.751074316 +0000 UTC m=+149.756779477" Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.751283 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm"] Dec 06 04:09:20 crc kubenswrapper[4718]: W1206 04:09:20.806793 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3c8d3d9_76af_41a6_92b0_eee699956823.slice/crio-3abee81d962f05d8ceeb6b4f5737912dd2d842b75da64cfae03f995cb4468514 WatchSource:0}: Error finding container 3abee81d962f05d8ceeb6b4f5737912dd2d842b75da64cfae03f995cb4468514: Status 404 returned error can't find the container with id 3abee81d962f05d8ceeb6b4f5737912dd2d842b75da64cfae03f995cb4468514 Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.831185 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:20 crc kubenswrapper[4718]: W1206 04:09:20.834082 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7dc6ff0_c499_46cf_bbfb_3f2c98e8eddf.slice/crio-e64831ad406c4d87e4d07724af37c72842665538712008969b173999a6a538ec WatchSource:0}: Error finding container e64831ad406c4d87e4d07724af37c72842665538712008969b173999a6a538ec: Status 404 returned error can't find the container with id e64831ad406c4d87e4d07724af37c72842665538712008969b173999a6a538ec Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.834211 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.334175665 +0000 UTC m=+150.339880836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.834340 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.835055 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.335038153 +0000 UTC m=+150.340743304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.842178 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.852875 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw"] Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.855034 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd"] Dec 06 04:09:20 crc kubenswrapper[4718]: W1206 04:09:20.897370 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71558f8b_d394_4546_9c82_c9ef5f10a4c4.slice/crio-e2736ce7c6ac9461f670133b612a01a54b18f2e19788ce053790331fade52c9a WatchSource:0}: Error finding container e2736ce7c6ac9461f670133b612a01a54b18f2e19788ce053790331fade52c9a: Status 404 returned error can't find the container with id e2736ce7c6ac9461f670133b612a01a54b18f2e19788ce053790331fade52c9a Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.935786 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.936113 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.436062648 +0000 UTC m=+150.441767809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.936714 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:20 crc kubenswrapper[4718]: E1206 04:09:20.937381 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.43737135 +0000 UTC m=+150.443076511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:20 crc kubenswrapper[4718]: I1206 04:09:20.983830 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.038832 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.039217 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.53918291 +0000 UTC m=+150.544888071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.086511 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.086559 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.132444 4718 patch_prober.go:28] interesting pod/apiserver-76f77b778f-q7kbb container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]log ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]etcd ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/generic-apiserver-start-informers ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/max-in-flight-filter ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 06 04:09:21 crc kubenswrapper[4718]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/project.openshift.io-projectcache ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/openshift.io-startinformers ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 06 04:09:21 crc kubenswrapper[4718]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 06 04:09:21 crc kubenswrapper[4718]: livez check failed Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.132503 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" podUID="7412230b-433d-4a8b-86d7-38cad91e6ecb" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.140791 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.141071 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.641061073 +0000 UTC m=+150.646766234 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.141403 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.142127 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.183000 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.227352 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:21 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:21 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:21 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.227633 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.242538 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.242684 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.742660956 +0000 UTC m=+150.748366117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.242903 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.244555 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.744544096 +0000 UTC m=+150.750249257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.343661 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.344012 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.8439976 +0000 UTC m=+150.849702751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.446918 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.447224 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:21.947212726 +0000 UTC m=+150.952917887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.547920 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.548352 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.048326983 +0000 UTC m=+151.054032134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.548650 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.548931 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.048923083 +0000 UTC m=+151.054628244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.649893 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.650791 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.150774875 +0000 UTC m=+151.156480036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.677606 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg" event={"ID":"8204ea9d-c0c7-4541-8e4e-a1b96303aa54","Type":"ContainerStarted","Data":"6b3e786137a80469224faeb6e98925d28fe93ba34fea03cd7cf12c58823abada"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.677653 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg" event={"ID":"8204ea9d-c0c7-4541-8e4e-a1b96303aa54","Type":"ContainerStarted","Data":"cde640af454704656d43557822d2f84538845a27991147c18c70b4de21ed0af9"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.689527 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" event={"ID":"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df","Type":"ContainerStarted","Data":"cee31a65d6e30116925ca270ec39c2d921ad776a4c7049254ec371ba9bef1ef0"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.695339 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" event={"ID":"982c800e-bebc-4214-a715-be8d91a05e83","Type":"ContainerStarted","Data":"000e8fb42a6c722080fc6ab1af9fe990d7a1c12eb694a112cbe338a3c161ba35"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.695390 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" event={"ID":"982c800e-bebc-4214-a715-be8d91a05e83","Type":"ContainerStarted","Data":"d1b14310520e69791cd3ed281c33c9227c67855dbc5c2065950bce3411e9cc74"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.696377 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.717750 4718 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vnhhd container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" start-of-body= Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.718253 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" podUID="982c800e-bebc-4214-a715-be8d91a05e83" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.737261 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" event={"ID":"e3a6e268-54a8-47d6-823e-fc3476ac9364","Type":"ContainerStarted","Data":"f56e8cf890db2406f9e9c6d1cd524323db2e0d676f5365a2e81139cbd1a920e7"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.742047 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" event={"ID":"cbccb659-e964-4146-a586-d71b85f000f0","Type":"ContainerStarted","Data":"4db38be5f03101109ede05ebb65501869b56dfa967f1a3e3c43e76e73b9bd269"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.742083 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" event={"ID":"cbccb659-e964-4146-a586-d71b85f000f0","Type":"ContainerStarted","Data":"723b44cc5d01bf6585ed72f19516cfd0011ea58a26d402869b0118e54da1f9b6"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.750954 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.751293 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v6j6g" event={"ID":"454877ff-cab8-4340-bfe7-1709bbd1227d","Type":"ContainerStarted","Data":"34cd8f1896871300f2d6e8e43456da84058442743bf20b582021d16f3631331c"} Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.751678 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.251664205 +0000 UTC m=+151.257369366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.753080 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" event={"ID":"73fa09b0-f084-4140-b014-febd78e2c2bf","Type":"ContainerStarted","Data":"6bf620641cef16470f4d1aa81d1c19eb7f814f6676f7f31597d07f1e36c29a92"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.754030 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" event={"ID":"d3c8d3d9-76af-41a6-92b0-eee699956823","Type":"ContainerStarted","Data":"3abee81d962f05d8ceeb6b4f5737912dd2d842b75da64cfae03f995cb4468514"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.755091 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" event={"ID":"3b0daeef-042f-4c26-b228-f879326917b0","Type":"ContainerStarted","Data":"80dcc669aaf712a7083fd6a323cce17cd033077021893156609520e10ae74fe0"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.758103 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.761015 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" event={"ID":"71558f8b-d394-4546-9c82-c9ef5f10a4c4","Type":"ContainerStarted","Data":"e2736ce7c6ac9461f670133b612a01a54b18f2e19788ce053790331fade52c9a"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.770733 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9e3f1db71520c1a8be2026d51f4332f832b3da10e5c358f56d63042d31fbbfe7"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.777706 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5z699" event={"ID":"e812f183-3419-4365-99f5-94f1e9f8abcb","Type":"ContainerStarted","Data":"8bd8a8c94c1d37c7ac961a2be37e9d84082c14be31f50f52719da5238f6c0356"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.785340 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" event={"ID":"b4aca9cc-4401-4463-9560-95dae620f6eb","Type":"ContainerStarted","Data":"a1d27170c1502b137bf0bec12b6f3111f6c1ce06288975a0ae0c52719ff59b84"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.788159 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" event={"ID":"64767824-fed2-496b-baa5-2c3ce0f8ee5b","Type":"ContainerStarted","Data":"d170d5f9f6f5bc45d825a98ed2b9eb74b95ff1afa65fec5e42aa50eab95b809e"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.809833 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"692a4e41789c5b08ac7f204b99b6cebc355f4cee808662e866599c698ed13b53"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.809883 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cf6e2c41aeb27c76e2fe35f09041341063b39d3810d8b2d5f5a024019a184956"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.810592 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.821014 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7jwsr" event={"ID":"76e6b6a8-fa61-4a22-a1c6-54a244a13696","Type":"ContainerStarted","Data":"537474c7cd58c397ef9adef91a161e5d824710c749a4a12e03d58d19eba25a84"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.823933 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" event={"ID":"7cf9c034-70ae-4c39-aca8-cc0a71a7a850","Type":"ContainerStarted","Data":"f75f826a2a92ee67cfad756e6bf2ed8bc5fd77af3616a072a73b91d4edee3997"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.823973 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" event={"ID":"7cf9c034-70ae-4c39-aca8-cc0a71a7a850","Type":"ContainerStarted","Data":"61971b8fc7dbf5e567909473dfaa8d03a1883147abb52ae24c3f83c836a17899"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.824685 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.830896 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" event={"ID":"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845","Type":"ContainerStarted","Data":"63d2d31fc3409be38027ef707421189ddbaaf9cc23ca04e758c4302a759fefec"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.833790 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" event={"ID":"96226f3a-94e6-4895-a9fc-662f07f8cdf8","Type":"ContainerStarted","Data":"f695d1825ea277f997f23c273dd1f5e4e6563a4379b0dbec52c9c1fd851ede3e"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.837114 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" event={"ID":"8b619ef6-1db6-4a92-9810-aa00e60f5bb8","Type":"ContainerStarted","Data":"598ac2b1eacbcd03d31974e2f66829c7b9f110af39f1fd283fc5b9dea325a418"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.837245 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" event={"ID":"8b619ef6-1db6-4a92-9810-aa00e60f5bb8","Type":"ContainerStarted","Data":"e538db296e72874a3fa7df816e7dec5be4f99fa0a1069fe1eb03fd6758acc392"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.843464 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" event={"ID":"ed46608d-0490-4b03-86ca-605ab4a57a93","Type":"ContainerStarted","Data":"5d7ff2b158058445eb3d264762aa6c047f6f546077341218e83b1b213ecdc506"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.844436 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" event={"ID":"a287c864-444a-41c1-bcfc-d11ed8086aef","Type":"ContainerStarted","Data":"c62739671daa12af83bcde96b08420d739dde3a6a7fd166dade66f4d111959cb"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.845455 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" event={"ID":"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82","Type":"ContainerStarted","Data":"0aea84451095014deea962a64aff471194143dabac211324da75b8288cfb60bc"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.845475 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" event={"ID":"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82","Type":"ContainerStarted","Data":"5801e2a88bc38774c83ef5960b7371c3f6baa34ac4d5571f74c7209173068cb0"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.846269 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mxjzh" event={"ID":"d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf","Type":"ContainerStarted","Data":"e64831ad406c4d87e4d07724af37c72842665538712008969b173999a6a538ec"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.846960 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" event={"ID":"daeca836-1e78-4156-9cbe-b065c3d2f59f","Type":"ContainerStarted","Data":"3fb4b8022440d30bddf343d032fe30e6ade7f285c84319f945d8c382cdc1515b"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.852904 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.854307 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.354286091 +0000 UTC m=+151.359991252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.862427 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-59sd7" event={"ID":"d22fc175-392e-4fdb-86c3-5d78793275d1","Type":"ContainerStarted","Data":"7b20dfbdd175a7db318b78a302c70c7748ea85e390269ed2f72a3962db030ca4"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.877893 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" event={"ID":"cb96422d-069d-48be-8a3d-da9a95a0e708","Type":"ContainerStarted","Data":"6aa423bb4ef9fd0250007978d710a85e17038a70471e574afaebf1627a09c268"} Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.882189 4718 patch_prober.go:28] interesting pod/downloads-7954f5f757-f84lw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.882263 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-f84lw" podUID="1308cc0b-85c6-49e9-a036-fc110f461f9c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.884532 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-69v98" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.899731 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-krdhn" Dec 06 04:09:21 crc kubenswrapper[4718]: I1206 04:09:21.955951 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:21 crc kubenswrapper[4718]: E1206 04:09:21.960820 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.460806383 +0000 UTC m=+151.466511534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.027192 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qsr" podStartSLOduration=132.027175754 podStartE2EDuration="2m12.027175754s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.025694307 +0000 UTC m=+151.031399468" watchObservedRunningTime="2025-12-06 04:09:22.027175754 +0000 UTC m=+151.032880915" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.058289 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" podStartSLOduration=132.058265663 podStartE2EDuration="2m12.058265663s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.0556589 +0000 UTC m=+151.061364061" watchObservedRunningTime="2025-12-06 04:09:22.058265663 +0000 UTC m=+151.063970824" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.060638 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.061025 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.561013242 +0000 UTC m=+151.566718403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.111277 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5kg75" podStartSLOduration=132.111257235 podStartE2EDuration="2m12.111257235s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.108890739 +0000 UTC m=+151.114595900" watchObservedRunningTime="2025-12-06 04:09:22.111257235 +0000 UTC m=+151.116962396" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.169868 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" podStartSLOduration=132.169854047 podStartE2EDuration="2m12.169854047s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.168789864 +0000 UTC m=+151.174495025" watchObservedRunningTime="2025-12-06 04:09:22.169854047 +0000 UTC m=+151.175559208" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.171199 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n76nm" podStartSLOduration=132.17119286 podStartE2EDuration="2m12.17119286s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.148639816 +0000 UTC m=+151.154344977" watchObservedRunningTime="2025-12-06 04:09:22.17119286 +0000 UTC m=+151.176898021" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.171770 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.172110 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.67209803 +0000 UTC m=+151.677803191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.226446 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:22 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:22 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:22 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.226693 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.236898 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-czh69" podStartSLOduration=132.23687812 podStartE2EDuration="2m12.23687812s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.188199346 +0000 UTC m=+151.193904507" watchObservedRunningTime="2025-12-06 04:09:22.23687812 +0000 UTC m=+151.242583281" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.273688 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.273999 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.773983592 +0000 UTC m=+151.779688753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.382861 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.383157 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.883144129 +0000 UTC m=+151.888849290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.385213 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" podStartSLOduration=132.385199974 podStartE2EDuration="2m12.385199974s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.3835225 +0000 UTC m=+151.389227671" watchObservedRunningTime="2025-12-06 04:09:22.385199974 +0000 UTC m=+151.390905135" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.402076 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-x2dzh" podStartSLOduration=132.402056966 podStartE2EDuration="2m12.402056966s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.35269717 +0000 UTC m=+151.358402351" watchObservedRunningTime="2025-12-06 04:09:22.402056966 +0000 UTC m=+151.407762127" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.483730 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.484491 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:22.984461573 +0000 UTC m=+151.990166734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.565698 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r4jb7" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.585320 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.585629 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:23.085615271 +0000 UTC m=+152.091320432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.587981 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" podStartSLOduration=131.587970467 podStartE2EDuration="2m11.587970467s" podCreationTimestamp="2025-12-06 04:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.464653536 +0000 UTC m=+151.470358697" watchObservedRunningTime="2025-12-06 04:09:22.587970467 +0000 UTC m=+151.593675628" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.686810 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.687111 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:23.187095971 +0000 UTC m=+152.192801122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.788920 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.789349 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:23.289332295 +0000 UTC m=+152.295037456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.889886 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.890269 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:23.390253597 +0000 UTC m=+152.395958758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.894827 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5z699" event={"ID":"e812f183-3419-4365-99f5-94f1e9f8abcb","Type":"ContainerStarted","Data":"d9ba838c0b28a937873c832345f77b91bcf790b0bdda9234aef4f878212796b2"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.897491 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" event={"ID":"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845","Type":"ContainerStarted","Data":"befec181c9ee2485645873c1b4022b35d584b42fbd0849eb58d281af10627b81"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.899789 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mxjzh" event={"ID":"d7dc6ff0-c499-46cf-bbfb-3f2c98e8eddf","Type":"ContainerStarted","Data":"2f0799d71aa33e121c846b685e449d888b8d2ccd321f8c035679f66518380a90"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.906174 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" event={"ID":"e3a6e268-54a8-47d6-823e-fc3476ac9364","Type":"ContainerStarted","Data":"18529af4d06ad607fac6bfa8565b157d77f1ff81356b9184dd133c579c4dbcb1"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.906207 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" event={"ID":"e3a6e268-54a8-47d6-823e-fc3476ac9364","Type":"ContainerStarted","Data":"f59678f9c18f8313f3fb2090a70c1d641344bbe744423dbe6e0099b1987e7db0"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.922028 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" event={"ID":"ed46608d-0490-4b03-86ca-605ab4a57a93","Type":"ContainerStarted","Data":"fc29955819e0f72221df33cf14c2fe0a808e45cf6c812b741e5cf5db31731327"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.924255 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" event={"ID":"0a1291c7-f265-4a3c-a994-3e8b3ab4e8df","Type":"ContainerStarted","Data":"6b60df0ffb022c5b7432c697fb74ed4e15c7a150f15fe807c08e602ee04cbe12"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.924880 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.926119 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" event={"ID":"3b0daeef-042f-4c26-b228-f879326917b0","Type":"ContainerStarted","Data":"522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.926634 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.935912 4718 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6kwkv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.935966 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" podUID="3b0daeef-042f-4c26-b228-f879326917b0" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.941054 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.942524 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-5z699" podStartSLOduration=132.942514165 podStartE2EDuration="2m12.942514165s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.940563583 +0000 UTC m=+151.946268744" watchObservedRunningTime="2025-12-06 04:09:22.942514165 +0000 UTC m=+151.948219326" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.963641 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" event={"ID":"58ce50cf-28a9-43a5-b4b3-bee39bbf7393","Type":"ContainerStarted","Data":"8509485055bbbc6ba10abacf46e8a9590af31195fcf96ca552269ea7c4462013"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.978634 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" podStartSLOduration=132.978607795 podStartE2EDuration="2m12.978607795s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:22.978538062 +0000 UTC m=+151.984243223" watchObservedRunningTime="2025-12-06 04:09:22.978607795 +0000 UTC m=+151.984312956" Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.995665 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqswv" event={"ID":"b4aca9cc-4401-4463-9560-95dae620f6eb","Type":"ContainerStarted","Data":"3cd1f35c2872f089ff61bc9ffbffda20f2155b8c77e27051775c23ef9d404217"} Dec 06 04:09:22 crc kubenswrapper[4718]: I1206 04:09:22.997374 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:22 crc kubenswrapper[4718]: E1206 04:09:22.997716 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:23.497701698 +0000 UTC m=+152.503406859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.011917 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" event={"ID":"a287c864-444a-41c1-bcfc-d11ed8086aef","Type":"ContainerStarted","Data":"a2128184d5e5590dda5147c4c67287cc0d2b91500feb456ff809979148dbf9b6"} Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.038452 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" event={"ID":"1adaef9f-42b9-4a1f-b1ad-6b4c7fcbeb82","Type":"ContainerStarted","Data":"a90f9e186ed461211b3e9417d3cf211e17354620b596b7a06d5ce367fb9f6c57"} Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.051697 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" podStartSLOduration=133.051675311 podStartE2EDuration="2m13.051675311s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.01987467 +0000 UTC m=+152.025579831" watchObservedRunningTime="2025-12-06 04:09:23.051675311 +0000 UTC m=+152.057380472" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.054217 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jn5bw" podStartSLOduration=133.054208363 podStartE2EDuration="2m13.054208363s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.049295895 +0000 UTC m=+152.055001076" watchObservedRunningTime="2025-12-06 04:09:23.054208363 +0000 UTC m=+152.059913524" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.096150 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg" event={"ID":"8204ea9d-c0c7-4541-8e4e-a1b96303aa54","Type":"ContainerStarted","Data":"e33ff37b4893ac67ff9f1b5d2b35809dd9d85825b951254f901c42134bdbbb65"} Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.099574 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:23 crc kubenswrapper[4718]: E1206 04:09:23.100899 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:23.600879452 +0000 UTC m=+152.606584613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.117507 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-59sd7" event={"ID":"d22fc175-392e-4fdb-86c3-5d78793275d1","Type":"ContainerStarted","Data":"f20b295e02666e6d0dbc66ea5951b75014a1db5990543f56e343e5cc6a7b8b6c"} Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.142613 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z7fbp" podStartSLOduration=133.142591952 podStartE2EDuration="2m13.142591952s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.113431135 +0000 UTC m=+152.119136296" watchObservedRunningTime="2025-12-06 04:09:23.142591952 +0000 UTC m=+152.148297113" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.144978 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" event={"ID":"71558f8b-d394-4546-9c82-c9ef5f10a4c4","Type":"ContainerStarted","Data":"1f54db09aaa83a413610bec17e65f4064cc9be11a24291e4334745ed9a694d62"} Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.145299 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.168093 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-gfknw" podStartSLOduration=132.16807737 podStartE2EDuration="2m12.16807737s" podCreationTimestamp="2025-12-06 04:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.145081192 +0000 UTC m=+152.150786353" watchObservedRunningTime="2025-12-06 04:09:23.16807737 +0000 UTC m=+152.173782531" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.168178 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-mxjzh" podStartSLOduration=8.168175103 podStartE2EDuration="8.168175103s" podCreationTimestamp="2025-12-06 04:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.166106497 +0000 UTC m=+152.171811658" watchObservedRunningTime="2025-12-06 04:09:23.168175103 +0000 UTC m=+152.173880264" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.179189 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" event={"ID":"daeca836-1e78-4156-9cbe-b065c3d2f59f","Type":"ContainerStarted","Data":"9086773ef55b58c5c89aa047d0c7bf7e87106841d8f2ad12157f130aedae200a"} Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.185570 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" podStartSLOduration=133.185550762 podStartE2EDuration="2m13.185550762s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.18363421 +0000 UTC m=+152.189339371" watchObservedRunningTime="2025-12-06 04:09:23.185550762 +0000 UTC m=+152.191255923" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.193711 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" event={"ID":"d3c8d3d9-76af-41a6-92b0-eee699956823","Type":"ContainerStarted","Data":"cd32173157538846079a4130cbe9fce97e1eeb763723ad62ac8d8bd208dbc0b9"} Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.200930 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:23 crc kubenswrapper[4718]: E1206 04:09:23.202064 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:23.702052961 +0000 UTC m=+152.707758122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.206583 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c4lmg" podStartSLOduration=133.206567657 podStartE2EDuration="2m13.206567657s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.204606243 +0000 UTC m=+152.210311404" watchObservedRunningTime="2025-12-06 04:09:23.206567657 +0000 UTC m=+152.212272818" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.218418 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7jwsr" event={"ID":"76e6b6a8-fa61-4a22-a1c6-54a244a13696","Type":"ContainerStarted","Data":"311a501e2e4f8d70d41d81c353186fb2c35051568935b1c2839e928d552bad45"} Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.225816 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-szgrx" podStartSLOduration=133.225797804 podStartE2EDuration="2m13.225797804s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.223133329 +0000 UTC m=+152.228838490" watchObservedRunningTime="2025-12-06 04:09:23.225797804 +0000 UTC m=+152.231502965" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.227309 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:23 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:23 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:23 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.227357 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.240528 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" event={"ID":"73fa09b0-f084-4140-b014-febd78e2c2bf","Type":"ContainerStarted","Data":"43841d9124d2b1f810f376b1851040301bc228cb1f206b564cf238e64086543b"} Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.272651 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vnhhd" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.304949 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:23 crc kubenswrapper[4718]: E1206 04:09:23.305559 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:23.805532665 +0000 UTC m=+152.811237906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.306215 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jzmpb" podStartSLOduration=133.306198497 podStartE2EDuration="2m13.306198497s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.305028329 +0000 UTC m=+152.310733500" watchObservedRunningTime="2025-12-06 04:09:23.306198497 +0000 UTC m=+152.311903658" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.337901 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" podStartSLOduration=133.337882895 podStartE2EDuration="2m13.337882895s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.337666847 +0000 UTC m=+152.343372008" watchObservedRunningTime="2025-12-06 04:09:23.337882895 +0000 UTC m=+152.343588056" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.408132 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rkfsp" podStartSLOduration=133.408115071 podStartE2EDuration="2m13.408115071s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.374824901 +0000 UTC m=+152.380530062" watchObservedRunningTime="2025-12-06 04:09:23.408115071 +0000 UTC m=+152.413820232" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.412578 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:23 crc kubenswrapper[4718]: E1206 04:09:23.416428 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:23.916411397 +0000 UTC m=+152.922116558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.431786 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h7222" podStartSLOduration=133.431765791 podStartE2EDuration="2m13.431765791s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:23.424062873 +0000 UTC m=+152.429768034" watchObservedRunningTime="2025-12-06 04:09:23.431765791 +0000 UTC m=+152.437470952" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.513942 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:23 crc kubenswrapper[4718]: E1206 04:09:23.514319 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.014302302 +0000 UTC m=+153.020007463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.615963 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:23 crc kubenswrapper[4718]: E1206 04:09:23.616456 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.116440411 +0000 UTC m=+153.122145572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.629092 4718 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.717430 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:23 crc kubenswrapper[4718]: E1206 04:09:23.717881 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.217845449 +0000 UTC m=+153.223550610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.819465 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:23 crc kubenswrapper[4718]: E1206 04:09:23.819879 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.319862376 +0000 UTC m=+153.325567537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:23 crc kubenswrapper[4718]: I1206 04:09:23.920630 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:23 crc kubenswrapper[4718]: E1206 04:09:23.920902 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.420886791 +0000 UTC m=+153.426591942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.022478 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:24 crc kubenswrapper[4718]: E1206 04:09:24.022849 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.522833286 +0000 UTC m=+153.528538447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.123673 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:24 crc kubenswrapper[4718]: E1206 04:09:24.124090 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.624058517 +0000 UTC m=+153.629763688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.225301 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.225583 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:24 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:24 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:24 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.225752 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:24 crc kubenswrapper[4718]: E1206 04:09:24.225679 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.725663981 +0000 UTC m=+153.731369142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.245878 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vp8rk" event={"ID":"daeca836-1e78-4156-9cbe-b065c3d2f59f","Type":"ContainerStarted","Data":"e4b966d591de90f4d07068c431daaea8dec49075a933a42edb3016a76a9c01b9"} Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.247217 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" event={"ID":"71558f8b-d394-4546-9c82-c9ef5f10a4c4","Type":"ContainerStarted","Data":"9f9c29f0b66e3463383228abf409d3db8815f8eb0efc240ce2ebc669f6004de4"} Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.248752 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-59sd7" event={"ID":"d22fc175-392e-4fdb-86c3-5d78793275d1","Type":"ContainerStarted","Data":"3933110d474c494c4010b9bdf880e20d2b9362b8053412134e48d11e5331d5aa"} Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.248782 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-59sd7" event={"ID":"d22fc175-392e-4fdb-86c3-5d78793275d1","Type":"ContainerStarted","Data":"745d8a0e2612ddb8974109d8172a2bf566567940b7fd9236eb4093c39f130693"} Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.249940 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7jwsr" event={"ID":"76e6b6a8-fa61-4a22-a1c6-54a244a13696","Type":"ContainerStarted","Data":"1b8f704d02ce188b5fda57a0cdb8ed8d9b7964a7407af1823b09e1909fc5fc67"} Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.250863 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.260105 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-29cng" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.265470 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.274304 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-7jwsr" podStartSLOduration=9.273712964 podStartE2EDuration="9.273712964s" podCreationTimestamp="2025-12-06 04:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:24.273564429 +0000 UTC m=+153.279269590" watchObservedRunningTime="2025-12-06 04:09:24.273712964 +0000 UTC m=+153.279418125" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.334850 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:24 crc kubenswrapper[4718]: E1206 04:09:24.335989 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.835964673 +0000 UTC m=+153.841669834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.437126 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:24 crc kubenswrapper[4718]: E1206 04:09:24.437479 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 04:09:24.937466443 +0000 UTC m=+153.943171604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sghf9" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.483623 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lnxxq"] Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.484781 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.486380 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.487326 4718 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-06T04:09:23.62912213Z","Handler":null,"Name":""} Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.489546 4718 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.489599 4718 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.493805 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lnxxq"] Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.538580 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.609564 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.640352 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-catalog-content\") pod \"certified-operators-lnxxq\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.640410 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.640441 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-utilities\") pod \"certified-operators-lnxxq\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.640520 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwctz\" (UniqueName: \"kubernetes.io/projected/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-kube-api-access-kwctz\") pod \"certified-operators-lnxxq\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.643300 4718 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.643335 4718 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.665946 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sghf9\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.687888 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwxl5"] Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.689976 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.693209 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.694413 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwxl5"] Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.741669 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwctz\" (UniqueName: \"kubernetes.io/projected/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-kube-api-access-kwctz\") pod \"certified-operators-lnxxq\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.741750 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-catalog-content\") pod \"certified-operators-lnxxq\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.741776 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-utilities\") pod \"certified-operators-lnxxq\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.742202 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-catalog-content\") pod \"certified-operators-lnxxq\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.742259 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-utilities\") pod \"certified-operators-lnxxq\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.763193 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwctz\" (UniqueName: \"kubernetes.io/projected/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-kube-api-access-kwctz\") pod \"certified-operators-lnxxq\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.766918 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.797915 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.842832 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-utilities\") pod \"community-operators-nwxl5\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.842930 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-catalog-content\") pod \"community-operators-nwxl5\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.842973 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7l2t\" (UniqueName: \"kubernetes.io/projected/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-kube-api-access-c7l2t\") pod \"community-operators-nwxl5\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.883827 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xg6xp"] Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.886833 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.904993 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xg6xp"] Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.944526 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-catalog-content\") pod \"community-operators-nwxl5\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.944567 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7l2t\" (UniqueName: \"kubernetes.io/projected/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-kube-api-access-c7l2t\") pod \"community-operators-nwxl5\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.944628 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-utilities\") pod \"community-operators-nwxl5\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.945513 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-utilities\") pod \"community-operators-nwxl5\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.945635 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-catalog-content\") pod \"community-operators-nwxl5\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:24 crc kubenswrapper[4718]: I1206 04:09:24.967527 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7l2t\" (UniqueName: \"kubernetes.io/projected/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-kube-api-access-c7l2t\") pod \"community-operators-nwxl5\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.002773 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.045364 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-utilities\") pod \"certified-operators-xg6xp\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.045693 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs9mk\" (UniqueName: \"kubernetes.io/projected/a8c4e5c8-91b8-432a-bf26-e59627d07754-kube-api-access-zs9mk\") pod \"certified-operators-xg6xp\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.045725 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-catalog-content\") pod \"certified-operators-xg6xp\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.064647 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sghf9"] Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.082191 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c4pxg"] Dec 06 04:09:25 crc kubenswrapper[4718]: W1206 04:09:25.082499 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb47eec75_b43c_486d_be96_dce4faceddbf.slice/crio-30d4fd3eefd2835a0e0684928307251fd68f1c35b0a22f12d87830925962f7b4 WatchSource:0}: Error finding container 30d4fd3eefd2835a0e0684928307251fd68f1c35b0a22f12d87830925962f7b4: Status 404 returned error can't find the container with id 30d4fd3eefd2835a0e0684928307251fd68f1c35b0a22f12d87830925962f7b4 Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.083226 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.094717 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lnxxq"] Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.104531 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c4pxg"] Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.147205 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-utilities\") pod \"certified-operators-xg6xp\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.147279 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs9mk\" (UniqueName: \"kubernetes.io/projected/a8c4e5c8-91b8-432a-bf26-e59627d07754-kube-api-access-zs9mk\") pod \"certified-operators-xg6xp\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.147305 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-catalog-content\") pod \"certified-operators-xg6xp\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.148048 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-utilities\") pod \"certified-operators-xg6xp\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.148307 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-catalog-content\") pod \"certified-operators-xg6xp\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.164798 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs9mk\" (UniqueName: \"kubernetes.io/projected/a8c4e5c8-91b8-432a-bf26-e59627d07754-kube-api-access-zs9mk\") pod \"certified-operators-xg6xp\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.205511 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.224560 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:25 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:25 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:25 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.224618 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.228830 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwxl5"] Dec 06 04:09:25 crc kubenswrapper[4718]: W1206 04:09:25.236925 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5ff337f_64c3_437b_a5cd_06dcf7cd96e8.slice/crio-59ebc6d50b095b89b78ca8e26d9c395b96dde70a2975a96ca137fae2a3ad4426 WatchSource:0}: Error finding container 59ebc6d50b095b89b78ca8e26d9c395b96dde70a2975a96ca137fae2a3ad4426: Status 404 returned error can't find the container with id 59ebc6d50b095b89b78ca8e26d9c395b96dde70a2975a96ca137fae2a3ad4426 Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.250075 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-catalog-content\") pod \"community-operators-c4pxg\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.250137 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6bvs\" (UniqueName: \"kubernetes.io/projected/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-kube-api-access-w6bvs\") pod \"community-operators-c4pxg\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.250219 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-utilities\") pod \"community-operators-c4pxg\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.258438 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnxxq" event={"ID":"989fbd15-c5fb-41b3-bdc8-2c5399b1424a","Type":"ContainerStarted","Data":"7739b998f987ee4bdb3a78b3c6acbe580cb4d62eaf457dc9657f2db827d70c61"} Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.260006 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" event={"ID":"b47eec75-b43c-486d-be96-dce4faceddbf","Type":"ContainerStarted","Data":"30d4fd3eefd2835a0e0684928307251fd68f1c35b0a22f12d87830925962f7b4"} Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.262788 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-59sd7" event={"ID":"d22fc175-392e-4fdb-86c3-5d78793275d1","Type":"ContainerStarted","Data":"a0b0f62478d8fc968ebdca28494cd038dbee435994e13a579675b03a5dca252e"} Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.264822 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwxl5" event={"ID":"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8","Type":"ContainerStarted","Data":"59ebc6d50b095b89b78ca8e26d9c395b96dde70a2975a96ca137fae2a3ad4426"} Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.292865 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-59sd7" podStartSLOduration=10.292846228 podStartE2EDuration="10.292846228s" podCreationTimestamp="2025-12-06 04:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:25.284322995 +0000 UTC m=+154.290028166" watchObservedRunningTime="2025-12-06 04:09:25.292846228 +0000 UTC m=+154.298551389" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.335059 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.353251 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-utilities\") pod \"community-operators-c4pxg\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.353519 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-catalog-content\") pod \"community-operators-c4pxg\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.353575 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6bvs\" (UniqueName: \"kubernetes.io/projected/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-kube-api-access-w6bvs\") pod \"community-operators-c4pxg\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.357007 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-utilities\") pod \"community-operators-c4pxg\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.357559 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-catalog-content\") pod \"community-operators-c4pxg\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.374440 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6bvs\" (UniqueName: \"kubernetes.io/projected/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-kube-api-access-w6bvs\") pod \"community-operators-c4pxg\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.407250 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.411392 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xg6xp"] Dec 06 04:09:25 crc kubenswrapper[4718]: I1206 04:09:25.600318 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c4pxg"] Dec 06 04:09:25 crc kubenswrapper[4718]: W1206 04:09:25.607800 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bbaf3bc_767a_4bfc_ad55_42c198b0b990.slice/crio-d07cb4f8beafef3e7cebcc3ca940feb51f69a2c032cef6044fb23263a23d6b81 WatchSource:0}: Error finding container d07cb4f8beafef3e7cebcc3ca940feb51f69a2c032cef6044fb23263a23d6b81: Status 404 returned error can't find the container with id d07cb4f8beafef3e7cebcc3ca940feb51f69a2c032cef6044fb23263a23d6b81 Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.094187 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.100312 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-q7kbb" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.227333 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:26 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:26 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:26 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.227396 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.269288 4718 generic.go:334] "Generic (PLEG): container finished" podID="f68d477b-1cc9-4ab9-8fa1-16fc30c9a845" containerID="befec181c9ee2485645873c1b4022b35d584b42fbd0849eb58d281af10627b81" exitCode=0 Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.269345 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" event={"ID":"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845","Type":"ContainerDied","Data":"befec181c9ee2485645873c1b4022b35d584b42fbd0849eb58d281af10627b81"} Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.271971 4718 generic.go:334] "Generic (PLEG): container finished" podID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerID="0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce" exitCode=0 Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.272033 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnxxq" event={"ID":"989fbd15-c5fb-41b3-bdc8-2c5399b1424a","Type":"ContainerDied","Data":"0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce"} Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.273483 4718 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.275686 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" event={"ID":"b47eec75-b43c-486d-be96-dce4faceddbf","Type":"ContainerStarted","Data":"39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36"} Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.275823 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.277965 4718 generic.go:334] "Generic (PLEG): container finished" podID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerID="0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f" exitCode=0 Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.278030 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pxg" event={"ID":"3bbaf3bc-767a-4bfc-ad55-42c198b0b990","Type":"ContainerDied","Data":"0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f"} Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.278053 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pxg" event={"ID":"3bbaf3bc-767a-4bfc-ad55-42c198b0b990","Type":"ContainerStarted","Data":"d07cb4f8beafef3e7cebcc3ca940feb51f69a2c032cef6044fb23263a23d6b81"} Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.279682 4718 generic.go:334] "Generic (PLEG): container finished" podID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerID="57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200" exitCode=0 Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.279726 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwxl5" event={"ID":"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8","Type":"ContainerDied","Data":"57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200"} Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.284179 4718 generic.go:334] "Generic (PLEG): container finished" podID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerID="660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531" exitCode=0 Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.284309 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg6xp" event={"ID":"a8c4e5c8-91b8-432a-bf26-e59627d07754","Type":"ContainerDied","Data":"660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531"} Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.284340 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg6xp" event={"ID":"a8c4e5c8-91b8-432a-bf26-e59627d07754","Type":"ContainerStarted","Data":"249db4e214fdbfcdce2aaf059a05ed45704c219d3831210bb855761c61ca9931"} Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.371687 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" podStartSLOduration=136.37166834 podStartE2EDuration="2m16.37166834s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:26.367979462 +0000 UTC m=+155.373684623" watchObservedRunningTime="2025-12-06 04:09:26.37166834 +0000 UTC m=+155.377373501" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.689420 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nkmtd"] Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.690477 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.692679 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.706432 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkmtd"] Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.772893 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-catalog-content\") pod \"redhat-marketplace-nkmtd\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.772977 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-utilities\") pod \"redhat-marketplace-nkmtd\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.773171 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5x8n\" (UniqueName: \"kubernetes.io/projected/a40acf1b-8c0d-4617-8f41-f48f097cce72-kube-api-access-j5x8n\") pod \"redhat-marketplace-nkmtd\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.874581 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-catalog-content\") pod \"redhat-marketplace-nkmtd\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.874681 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-utilities\") pod \"redhat-marketplace-nkmtd\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.874767 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5x8n\" (UniqueName: \"kubernetes.io/projected/a40acf1b-8c0d-4617-8f41-f48f097cce72-kube-api-access-j5x8n\") pod \"redhat-marketplace-nkmtd\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.875743 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-catalog-content\") pod \"redhat-marketplace-nkmtd\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.875763 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-utilities\") pod \"redhat-marketplace-nkmtd\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:26 crc kubenswrapper[4718]: I1206 04:09:26.908405 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5x8n\" (UniqueName: \"kubernetes.io/projected/a40acf1b-8c0d-4617-8f41-f48f097cce72-kube-api-access-j5x8n\") pod \"redhat-marketplace-nkmtd\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.008001 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.086031 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdnb"] Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.087383 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.103465 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdnb"] Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.190921 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-catalog-content\") pod \"redhat-marketplace-bhdnb\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.190973 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-utilities\") pod \"redhat-marketplace-bhdnb\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.191038 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjvdt\" (UniqueName: \"kubernetes.io/projected/fbf10d92-c15e-4175-ae97-42e90be68c10-kube-api-access-fjvdt\") pod \"redhat-marketplace-bhdnb\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.227731 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:27 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:27 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:27 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.227791 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.291808 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjvdt\" (UniqueName: \"kubernetes.io/projected/fbf10d92-c15e-4175-ae97-42e90be68c10-kube-api-access-fjvdt\") pod \"redhat-marketplace-bhdnb\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.291903 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-catalog-content\") pod \"redhat-marketplace-bhdnb\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.291929 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-utilities\") pod \"redhat-marketplace-bhdnb\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.292829 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-utilities\") pod \"redhat-marketplace-bhdnb\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.293047 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-catalog-content\") pod \"redhat-marketplace-bhdnb\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.331103 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjvdt\" (UniqueName: \"kubernetes.io/projected/fbf10d92-c15e-4175-ae97-42e90be68c10-kube-api-access-fjvdt\") pod \"redhat-marketplace-bhdnb\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.391814 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkmtd"] Dec 06 04:09:27 crc kubenswrapper[4718]: W1206 04:09:27.398107 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda40acf1b_8c0d_4617_8f41_f48f097cce72.slice/crio-6b1a4ae1e4d4cdc0ac646cfd3817def4f7a997369b19e324d2df1865ece105b0 WatchSource:0}: Error finding container 6b1a4ae1e4d4cdc0ac646cfd3817def4f7a997369b19e324d2df1865ece105b0: Status 404 returned error can't find the container with id 6b1a4ae1e4d4cdc0ac646cfd3817def4f7a997369b19e324d2df1865ece105b0 Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.437731 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.471128 4718 patch_prober.go:28] interesting pod/downloads-7954f5f757-f84lw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.471156 4718 patch_prober.go:28] interesting pod/downloads-7954f5f757-f84lw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.471186 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-f84lw" podUID="1308cc0b-85c6-49e9-a036-fc110f461f9c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.471249 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-f84lw" podUID="1308cc0b-85c6-49e9-a036-fc110f461f9c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.503345 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.503405 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.506430 4718 patch_prober.go:28] interesting pod/console-f9d7485db-v6j6g container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.506537 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-v6j6g" podUID="454877ff-cab8-4340-bfe7-1709bbd1227d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.648978 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.687105 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-47mm6"] Dec 06 04:09:27 crc kubenswrapper[4718]: E1206 04:09:27.687362 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f68d477b-1cc9-4ab9-8fa1-16fc30c9a845" containerName="collect-profiles" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.687380 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f68d477b-1cc9-4ab9-8fa1-16fc30c9a845" containerName="collect-profiles" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.687472 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f68d477b-1cc9-4ab9-8fa1-16fc30c9a845" containerName="collect-profiles" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.688149 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.697278 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.700970 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-47mm6"] Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.765711 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdnb"] Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.808651 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-config-volume\") pod \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.808800 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-secret-volume\") pod \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.808830 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lncxn\" (UniqueName: \"kubernetes.io/projected/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-kube-api-access-lncxn\") pod \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\" (UID: \"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845\") " Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.809069 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-utilities\") pod \"redhat-operators-47mm6\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.809108 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zqwb\" (UniqueName: \"kubernetes.io/projected/b5790280-687b-4e83-addd-504672823e57-kube-api-access-8zqwb\") pod \"redhat-operators-47mm6\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.809158 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-catalog-content\") pod \"redhat-operators-47mm6\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.813658 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-config-volume" (OuterVolumeSpecName: "config-volume") pod "f68d477b-1cc9-4ab9-8fa1-16fc30c9a845" (UID: "f68d477b-1cc9-4ab9-8fa1-16fc30c9a845"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.833421 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-kube-api-access-lncxn" (OuterVolumeSpecName: "kube-api-access-lncxn") pod "f68d477b-1cc9-4ab9-8fa1-16fc30c9a845" (UID: "f68d477b-1cc9-4ab9-8fa1-16fc30c9a845"). InnerVolumeSpecName "kube-api-access-lncxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.833517 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f68d477b-1cc9-4ab9-8fa1-16fc30c9a845" (UID: "f68d477b-1cc9-4ab9-8fa1-16fc30c9a845"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.880331 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.880397 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.912282 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-utilities\") pod \"redhat-operators-47mm6\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.912342 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zqwb\" (UniqueName: \"kubernetes.io/projected/b5790280-687b-4e83-addd-504672823e57-kube-api-access-8zqwb\") pod \"redhat-operators-47mm6\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.912381 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-catalog-content\") pod \"redhat-operators-47mm6\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.912457 4718 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.912484 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lncxn\" (UniqueName: \"kubernetes.io/projected/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-kube-api-access-lncxn\") on node \"crc\" DevicePath \"\"" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.912493 4718 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.915314 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-catalog-content\") pod \"redhat-operators-47mm6\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.916722 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-utilities\") pod \"redhat-operators-47mm6\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.931589 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zqwb\" (UniqueName: \"kubernetes.io/projected/b5790280-687b-4e83-addd-504672823e57-kube-api-access-8zqwb\") pod \"redhat-operators-47mm6\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.953592 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.955315 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.958123 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.958365 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 06 04:09:27 crc kubenswrapper[4718]: I1206 04:09:27.962168 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.057894 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.082798 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s22px"] Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.083884 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.092203 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s22px"] Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.115219 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14f60bde-3edc-4502-b3c6-d922be8b22f2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"14f60bde-3edc-4502-b3c6-d922be8b22f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.115311 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14f60bde-3edc-4502-b3c6-d922be8b22f2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"14f60bde-3edc-4502-b3c6-d922be8b22f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.216340 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpx9f\" (UniqueName: \"kubernetes.io/projected/360c8eaf-4d22-4e90-acbd-91d78cb1249b-kube-api-access-vpx9f\") pod \"redhat-operators-s22px\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.216399 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14f60bde-3edc-4502-b3c6-d922be8b22f2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"14f60bde-3edc-4502-b3c6-d922be8b22f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.216438 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-utilities\") pod \"redhat-operators-s22px\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.216505 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-catalog-content\") pod \"redhat-operators-s22px\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.216527 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14f60bde-3edc-4502-b3c6-d922be8b22f2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"14f60bde-3edc-4502-b3c6-d922be8b22f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.217001 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14f60bde-3edc-4502-b3c6-d922be8b22f2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"14f60bde-3edc-4502-b3c6-d922be8b22f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.222575 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.228727 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:28 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:28 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:28 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.228811 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.248812 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14f60bde-3edc-4502-b3c6-d922be8b22f2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"14f60bde-3edc-4502-b3c6-d922be8b22f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.273685 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.307440 4718 generic.go:334] "Generic (PLEG): container finished" podID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerID="a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468" exitCode=0 Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.307539 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkmtd" event={"ID":"a40acf1b-8c0d-4617-8f41-f48f097cce72","Type":"ContainerDied","Data":"a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468"} Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.307564 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkmtd" event={"ID":"a40acf1b-8c0d-4617-8f41-f48f097cce72","Type":"ContainerStarted","Data":"6b1a4ae1e4d4cdc0ac646cfd3817def4f7a997369b19e324d2df1865ece105b0"} Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.311628 4718 generic.go:334] "Generic (PLEG): container finished" podID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerID="43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971" exitCode=0 Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.312368 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdnb" event={"ID":"fbf10d92-c15e-4175-ae97-42e90be68c10","Type":"ContainerDied","Data":"43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971"} Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.312399 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdnb" event={"ID":"fbf10d92-c15e-4175-ae97-42e90be68c10","Type":"ContainerStarted","Data":"f4ec6c7ae75bad2987a9212f021079783050b2c520bb4a990a0296f415b541a9"} Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.317783 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-utilities\") pod \"redhat-operators-s22px\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.317861 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-catalog-content\") pod \"redhat-operators-s22px\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.318022 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpx9f\" (UniqueName: \"kubernetes.io/projected/360c8eaf-4d22-4e90-acbd-91d78cb1249b-kube-api-access-vpx9f\") pod \"redhat-operators-s22px\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.320075 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-utilities\") pod \"redhat-operators-s22px\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.320671 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-catalog-content\") pod \"redhat-operators-s22px\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.321376 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.321825 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm" event={"ID":"f68d477b-1cc9-4ab9-8fa1-16fc30c9a845","Type":"ContainerDied","Data":"63d2d31fc3409be38027ef707421189ddbaaf9cc23ca04e758c4302a759fefec"} Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.321880 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63d2d31fc3409be38027ef707421189ddbaaf9cc23ca04e758c4302a759fefec" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.334087 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpx9f\" (UniqueName: \"kubernetes.io/projected/360c8eaf-4d22-4e90-acbd-91d78cb1249b-kube-api-access-vpx9f\") pod \"redhat-operators-s22px\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.435586 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.471565 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-47mm6"] Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.763813 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s22px"] Dec 06 04:09:28 crc kubenswrapper[4718]: I1206 04:09:28.783208 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 04:09:29 crc kubenswrapper[4718]: I1206 04:09:29.225917 4718 patch_prober.go:28] interesting pod/router-default-5444994796-w4fdl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 04:09:29 crc kubenswrapper[4718]: [-]has-synced failed: reason withheld Dec 06 04:09:29 crc kubenswrapper[4718]: [+]process-running ok Dec 06 04:09:29 crc kubenswrapper[4718]: healthz check failed Dec 06 04:09:29 crc kubenswrapper[4718]: I1206 04:09:29.226164 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w4fdl" podUID="8ae3fa3c-0ea2-40a0-b053-6a8f6d33ecb5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 04:09:29 crc kubenswrapper[4718]: I1206 04:09:29.342000 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"14f60bde-3edc-4502-b3c6-d922be8b22f2","Type":"ContainerStarted","Data":"f23423be710e5bbe604b7bca1acc5a52198a277149dad8b2038cc7aa23b1dc31"} Dec 06 04:09:29 crc kubenswrapper[4718]: I1206 04:09:29.347422 4718 generic.go:334] "Generic (PLEG): container finished" podID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerID="7a429892afc695df0cae3397cf4770f0591564d67de90a8ab8d8e988c955484d" exitCode=0 Dec 06 04:09:29 crc kubenswrapper[4718]: I1206 04:09:29.348216 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s22px" event={"ID":"360c8eaf-4d22-4e90-acbd-91d78cb1249b","Type":"ContainerDied","Data":"7a429892afc695df0cae3397cf4770f0591564d67de90a8ab8d8e988c955484d"} Dec 06 04:09:29 crc kubenswrapper[4718]: I1206 04:09:29.348257 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s22px" event={"ID":"360c8eaf-4d22-4e90-acbd-91d78cb1249b","Type":"ContainerStarted","Data":"7c4fd01697de5a4e7dd940dc6917a69b7eb586724d8f15a1772e750255d33308"} Dec 06 04:09:29 crc kubenswrapper[4718]: I1206 04:09:29.363329 4718 generic.go:334] "Generic (PLEG): container finished" podID="b5790280-687b-4e83-addd-504672823e57" containerID="691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe" exitCode=0 Dec 06 04:09:29 crc kubenswrapper[4718]: I1206 04:09:29.363373 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47mm6" event={"ID":"b5790280-687b-4e83-addd-504672823e57","Type":"ContainerDied","Data":"691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe"} Dec 06 04:09:29 crc kubenswrapper[4718]: I1206 04:09:29.365653 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47mm6" event={"ID":"b5790280-687b-4e83-addd-504672823e57","Type":"ContainerStarted","Data":"402586121dcfe4de78cc88c9b6afc12b86612d8389a63e2257f3a0b830cc11b3"} Dec 06 04:09:30 crc kubenswrapper[4718]: I1206 04:09:30.227643 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:30 crc kubenswrapper[4718]: I1206 04:09:30.235555 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-w4fdl" Dec 06 04:09:30 crc kubenswrapper[4718]: I1206 04:09:30.377162 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"14f60bde-3edc-4502-b3c6-d922be8b22f2","Type":"ContainerStarted","Data":"0ba7234cb10619fb471d3ce9879edd0c3c47dbeaf60f04c1b1f615f43b78318f"} Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.246894 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.246876832 podStartE2EDuration="4.246876832s" podCreationTimestamp="2025-12-06 04:09:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:09:30.39283664 +0000 UTC m=+159.398541801" watchObservedRunningTime="2025-12-06 04:09:31.246876832 +0000 UTC m=+160.252581993" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.248606 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.249287 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.251222 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.251439 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.261767 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.377628 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.377705 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.388306 4718 generic.go:334] "Generic (PLEG): container finished" podID="14f60bde-3edc-4502-b3c6-d922be8b22f2" containerID="0ba7234cb10619fb471d3ce9879edd0c3c47dbeaf60f04c1b1f615f43b78318f" exitCode=0 Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.388355 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"14f60bde-3edc-4502-b3c6-d922be8b22f2","Type":"ContainerDied","Data":"0ba7234cb10619fb471d3ce9879edd0c3c47dbeaf60f04c1b1f615f43b78318f"} Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.478319 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.478409 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.478509 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.498721 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:09:31 crc kubenswrapper[4718]: I1206 04:09:31.575698 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:09:32 crc kubenswrapper[4718]: I1206 04:09:32.085499 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 04:09:32 crc kubenswrapper[4718]: I1206 04:09:32.397828 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c19fb5f4-5d28-42d7-8e51-eec530a5ce39","Type":"ContainerStarted","Data":"30b7ed707c8320d965ef16362ce503f5c0ed4bdaf34127626f24bb2440f3cedb"} Dec 06 04:09:32 crc kubenswrapper[4718]: I1206 04:09:32.637428 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:32 crc kubenswrapper[4718]: I1206 04:09:32.801358 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14f60bde-3edc-4502-b3c6-d922be8b22f2-kubelet-dir\") pod \"14f60bde-3edc-4502-b3c6-d922be8b22f2\" (UID: \"14f60bde-3edc-4502-b3c6-d922be8b22f2\") " Dec 06 04:09:32 crc kubenswrapper[4718]: I1206 04:09:32.801496 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14f60bde-3edc-4502-b3c6-d922be8b22f2-kube-api-access\") pod \"14f60bde-3edc-4502-b3c6-d922be8b22f2\" (UID: \"14f60bde-3edc-4502-b3c6-d922be8b22f2\") " Dec 06 04:09:32 crc kubenswrapper[4718]: I1206 04:09:32.801490 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14f60bde-3edc-4502-b3c6-d922be8b22f2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "14f60bde-3edc-4502-b3c6-d922be8b22f2" (UID: "14f60bde-3edc-4502-b3c6-d922be8b22f2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:09:32 crc kubenswrapper[4718]: I1206 04:09:32.801717 4718 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14f60bde-3edc-4502-b3c6-d922be8b22f2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:09:32 crc kubenswrapper[4718]: I1206 04:09:32.808435 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f60bde-3edc-4502-b3c6-d922be8b22f2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "14f60bde-3edc-4502-b3c6-d922be8b22f2" (UID: "14f60bde-3edc-4502-b3c6-d922be8b22f2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:09:32 crc kubenswrapper[4718]: I1206 04:09:32.902706 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14f60bde-3edc-4502-b3c6-d922be8b22f2-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 04:09:33 crc kubenswrapper[4718]: I1206 04:09:33.210884 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:09:33 crc kubenswrapper[4718]: I1206 04:09:33.214035 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38892dbb-af9b-4483-bfe6-c1ed160d356c-metrics-certs\") pod \"network-metrics-daemon-mhz2w\" (UID: \"38892dbb-af9b-4483-bfe6-c1ed160d356c\") " pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:09:33 crc kubenswrapper[4718]: I1206 04:09:33.243445 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-7jwsr" Dec 06 04:09:33 crc kubenswrapper[4718]: I1206 04:09:33.409633 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"14f60bde-3edc-4502-b3c6-d922be8b22f2","Type":"ContainerDied","Data":"f23423be710e5bbe604b7bca1acc5a52198a277149dad8b2038cc7aa23b1dc31"} Dec 06 04:09:33 crc kubenswrapper[4718]: I1206 04:09:33.409695 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 04:09:33 crc kubenswrapper[4718]: I1206 04:09:33.409698 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f23423be710e5bbe604b7bca1acc5a52198a277149dad8b2038cc7aa23b1dc31" Dec 06 04:09:33 crc kubenswrapper[4718]: I1206 04:09:33.477352 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mhz2w" Dec 06 04:09:33 crc kubenswrapper[4718]: I1206 04:09:33.866529 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mhz2w"] Dec 06 04:09:33 crc kubenswrapper[4718]: W1206 04:09:33.888423 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38892dbb_af9b_4483_bfe6_c1ed160d356c.slice/crio-cb608da1d9c891d5e71afc2fff8628e9c3090fbe37e6b3cdbc4a3a4ced76d19f WatchSource:0}: Error finding container cb608da1d9c891d5e71afc2fff8628e9c3090fbe37e6b3cdbc4a3a4ced76d19f: Status 404 returned error can't find the container with id cb608da1d9c891d5e71afc2fff8628e9c3090fbe37e6b3cdbc4a3a4ced76d19f Dec 06 04:09:34 crc kubenswrapper[4718]: I1206 04:09:34.421194 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c19fb5f4-5d28-42d7-8e51-eec530a5ce39","Type":"ContainerStarted","Data":"255f49a7b7933c636933c5a68fb677e754f6ded8a042c652f7434fba9fa9ee6f"} Dec 06 04:09:34 crc kubenswrapper[4718]: I1206 04:09:34.445842 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" event={"ID":"38892dbb-af9b-4483-bfe6-c1ed160d356c","Type":"ContainerStarted","Data":"cb608da1d9c891d5e71afc2fff8628e9c3090fbe37e6b3cdbc4a3a4ced76d19f"} Dec 06 04:09:35 crc kubenswrapper[4718]: I1206 04:09:35.453214 4718 generic.go:334] "Generic (PLEG): container finished" podID="c19fb5f4-5d28-42d7-8e51-eec530a5ce39" containerID="255f49a7b7933c636933c5a68fb677e754f6ded8a042c652f7434fba9fa9ee6f" exitCode=0 Dec 06 04:09:35 crc kubenswrapper[4718]: I1206 04:09:35.453597 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c19fb5f4-5d28-42d7-8e51-eec530a5ce39","Type":"ContainerDied","Data":"255f49a7b7933c636933c5a68fb677e754f6ded8a042c652f7434fba9fa9ee6f"} Dec 06 04:09:35 crc kubenswrapper[4718]: I1206 04:09:35.458136 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" event={"ID":"38892dbb-af9b-4483-bfe6-c1ed160d356c","Type":"ContainerStarted","Data":"5ed9163f8e6e0c1e36b2a5c743e681a0e32ca387196d635712a0c2e87d2fd2c0"} Dec 06 04:09:37 crc kubenswrapper[4718]: I1206 04:09:37.476910 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-f84lw" Dec 06 04:09:37 crc kubenswrapper[4718]: I1206 04:09:37.507811 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:37 crc kubenswrapper[4718]: I1206 04:09:37.511980 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-v6j6g" Dec 06 04:09:44 crc kubenswrapper[4718]: I1206 04:09:44.775608 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:09:57 crc kubenswrapper[4718]: I1206 04:09:57.557975 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 04:09:57 crc kubenswrapper[4718]: I1206 04:09:57.877025 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:09:57 crc kubenswrapper[4718]: I1206 04:09:57.877086 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:09:58 crc kubenswrapper[4718]: I1206 04:09:58.588566 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5mv6" Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.125273 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.247504 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kubelet-dir\") pod \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\" (UID: \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\") " Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.247629 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kube-api-access\") pod \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\" (UID: \"c19fb5f4-5d28-42d7-8e51-eec530a5ce39\") " Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.247685 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c19fb5f4-5d28-42d7-8e51-eec530a5ce39" (UID: "c19fb5f4-5d28-42d7-8e51-eec530a5ce39"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.248281 4718 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.253961 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c19fb5f4-5d28-42d7-8e51-eec530a5ce39" (UID: "c19fb5f4-5d28-42d7-8e51-eec530a5ce39"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.349957 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19fb5f4-5d28-42d7-8e51-eec530a5ce39-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.620899 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c19fb5f4-5d28-42d7-8e51-eec530a5ce39","Type":"ContainerDied","Data":"30b7ed707c8320d965ef16362ce503f5c0ed4bdaf34127626f24bb2440f3cedb"} Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.620956 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30b7ed707c8320d965ef16362ce503f5c0ed4bdaf34127626f24bb2440f3cedb" Dec 06 04:10:02 crc kubenswrapper[4718]: I1206 04:10:02.620989 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.442662 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 04:10:04 crc kubenswrapper[4718]: E1206 04:10:04.443070 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19fb5f4-5d28-42d7-8e51-eec530a5ce39" containerName="pruner" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.443082 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19fb5f4-5d28-42d7-8e51-eec530a5ce39" containerName="pruner" Dec 06 04:10:04 crc kubenswrapper[4718]: E1206 04:10:04.443098 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f60bde-3edc-4502-b3c6-d922be8b22f2" containerName="pruner" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.443105 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f60bde-3edc-4502-b3c6-d922be8b22f2" containerName="pruner" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.443218 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19fb5f4-5d28-42d7-8e51-eec530a5ce39" containerName="pruner" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.443259 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f60bde-3edc-4502-b3c6-d922be8b22f2" containerName="pruner" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.444125 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.449123 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.451115 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.452822 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.496119 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94e3d13a-7d03-473d-81e7-611590efce2d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"94e3d13a-7d03-473d-81e7-611590efce2d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.496176 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/94e3d13a-7d03-473d-81e7-611590efce2d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"94e3d13a-7d03-473d-81e7-611590efce2d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.597842 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94e3d13a-7d03-473d-81e7-611590efce2d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"94e3d13a-7d03-473d-81e7-611590efce2d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.597910 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/94e3d13a-7d03-473d-81e7-611590efce2d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"94e3d13a-7d03-473d-81e7-611590efce2d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.598000 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/94e3d13a-7d03-473d-81e7-611590efce2d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"94e3d13a-7d03-473d-81e7-611590efce2d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.614994 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94e3d13a-7d03-473d-81e7-611590efce2d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"94e3d13a-7d03-473d-81e7-611590efce2d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:04 crc kubenswrapper[4718]: I1206 04:10:04.761127 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:07 crc kubenswrapper[4718]: E1206 04:10:07.816984 4718 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 04:10:07 crc kubenswrapper[4718]: E1206 04:10:07.817458 4718 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fjvdt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bhdnb_openshift-marketplace(fbf10d92-c15e-4175-ae97-42e90be68c10): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 04:10:07 crc kubenswrapper[4718]: E1206 04:10:07.818637 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bhdnb" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" Dec 06 04:10:08 crc kubenswrapper[4718]: E1206 04:10:08.191278 4718 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 04:10:08 crc kubenswrapper[4718]: E1206 04:10:08.191422 4718 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j5x8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-nkmtd_openshift-marketplace(a40acf1b-8c0d-4617-8f41-f48f097cce72): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 04:10:08 crc kubenswrapper[4718]: E1206 04:10:08.193464 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-nkmtd" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.448029 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.449618 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.453741 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bde102c1-6414-4a3a-b463-8f57a44257c7-kube-api-access\") pod \"installer-9-crc\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.453858 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.453939 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-var-lock\") pod \"installer-9-crc\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.459867 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.555872 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-var-lock\") pod \"installer-9-crc\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.555973 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bde102c1-6414-4a3a-b463-8f57a44257c7-kube-api-access\") pod \"installer-9-crc\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.555996 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.556034 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-var-lock\") pod \"installer-9-crc\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.556088 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.587495 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bde102c1-6414-4a3a-b463-8f57a44257c7-kube-api-access\") pod \"installer-9-crc\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:09 crc kubenswrapper[4718]: I1206 04:10:09.770567 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:10:12 crc kubenswrapper[4718]: E1206 04:10:12.440281 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bhdnb" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" Dec 06 04:10:12 crc kubenswrapper[4718]: E1206 04:10:12.440297 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-nkmtd" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" Dec 06 04:10:13 crc kubenswrapper[4718]: E1206 04:10:13.810388 4718 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 04:10:13 crc kubenswrapper[4718]: E1206 04:10:13.810905 4718 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zs9mk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-xg6xp_openshift-marketplace(a8c4e5c8-91b8-432a-bf26-e59627d07754): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 04:10:13 crc kubenswrapper[4718]: E1206 04:10:13.812144 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-xg6xp" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" Dec 06 04:10:16 crc kubenswrapper[4718]: E1206 04:10:16.572901 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-xg6xp" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" Dec 06 04:10:20 crc kubenswrapper[4718]: E1206 04:10:20.665931 4718 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 04:10:20 crc kubenswrapper[4718]: E1206 04:10:20.666323 4718 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vpx9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-s22px_openshift-marketplace(360c8eaf-4d22-4e90-acbd-91d78cb1249b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 04:10:20 crc kubenswrapper[4718]: E1206 04:10:20.667520 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-s22px" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" Dec 06 04:10:20 crc kubenswrapper[4718]: E1206 04:10:20.706496 4718 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 04:10:20 crc kubenswrapper[4718]: E1206 04:10:20.706717 4718 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8zqwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-47mm6_openshift-marketplace(b5790280-687b-4e83-addd-504672823e57): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 04:10:20 crc kubenswrapper[4718]: E1206 04:10:20.708660 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-47mm6" podUID="b5790280-687b-4e83-addd-504672823e57" Dec 06 04:10:20 crc kubenswrapper[4718]: E1206 04:10:20.783737 4718 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 04:10:20 crc kubenswrapper[4718]: E1206 04:10:20.783883 4718 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kwctz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-lnxxq_openshift-marketplace(989fbd15-c5fb-41b3-bdc8-2c5399b1424a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 04:10:20 crc kubenswrapper[4718]: E1206 04:10:20.785362 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-lnxxq" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.226809 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-lnxxq" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.227177 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-47mm6" podUID="b5790280-687b-4e83-addd-504672823e57" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.227324 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-s22px" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.318611 4718 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.318998 4718 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w6bvs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-c4pxg_openshift-marketplace(3bbaf3bc-767a-4bfc-ad55-42c198b0b990): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.320316 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-c4pxg" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.350545 4718 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.350692 4718 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c7l2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-nwxl5_openshift-marketplace(b5ff337f-64c3-437b-a5cd-06dcf7cd96e8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.352032 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-nwxl5" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" Dec 06 04:10:22 crc kubenswrapper[4718]: I1206 04:10:22.642110 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 04:10:22 crc kubenswrapper[4718]: W1206 04:10:22.651885 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod94e3d13a_7d03_473d_81e7_611590efce2d.slice/crio-e565c8e716b71b5620f8634a04a797ddfe2c526aa6ec05ea5dd995039c5efc86 WatchSource:0}: Error finding container e565c8e716b71b5620f8634a04a797ddfe2c526aa6ec05ea5dd995039c5efc86: Status 404 returned error can't find the container with id e565c8e716b71b5620f8634a04a797ddfe2c526aa6ec05ea5dd995039c5efc86 Dec 06 04:10:22 crc kubenswrapper[4718]: I1206 04:10:22.704275 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 04:10:22 crc kubenswrapper[4718]: I1206 04:10:22.741859 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mhz2w" event={"ID":"38892dbb-af9b-4483-bfe6-c1ed160d356c","Type":"ContainerStarted","Data":"1b96542b0afb4a8c90739becf8e2274ce9ed7453303fbfe1f34d74d0ed8eea70"} Dec 06 04:10:22 crc kubenswrapper[4718]: I1206 04:10:22.746918 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"94e3d13a-7d03-473d-81e7-611590efce2d","Type":"ContainerStarted","Data":"e565c8e716b71b5620f8634a04a797ddfe2c526aa6ec05ea5dd995039c5efc86"} Dec 06 04:10:22 crc kubenswrapper[4718]: I1206 04:10:22.748961 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bde102c1-6414-4a3a-b463-8f57a44257c7","Type":"ContainerStarted","Data":"2874a166a5b920c5e1a33551548b16c0bae0e5eb166b6f4d1e464878574ab026"} Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.750957 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-nwxl5" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" Dec 06 04:10:22 crc kubenswrapper[4718]: E1206 04:10:22.751457 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-c4pxg" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" Dec 06 04:10:22 crc kubenswrapper[4718]: I1206 04:10:22.759317 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-mhz2w" podStartSLOduration=192.759298783 podStartE2EDuration="3m12.759298783s" podCreationTimestamp="2025-12-06 04:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:10:22.755799075 +0000 UTC m=+211.761504256" watchObservedRunningTime="2025-12-06 04:10:22.759298783 +0000 UTC m=+211.765003944" Dec 06 04:10:23 crc kubenswrapper[4718]: I1206 04:10:23.760292 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bde102c1-6414-4a3a-b463-8f57a44257c7","Type":"ContainerStarted","Data":"c60f02bc09933f7fee870bb3350a4550280b59696c6dbd814755940d1cf395e6"} Dec 06 04:10:23 crc kubenswrapper[4718]: I1206 04:10:23.763782 4718 generic.go:334] "Generic (PLEG): container finished" podID="94e3d13a-7d03-473d-81e7-611590efce2d" containerID="8d2bbede898d1fc39452b1b49eefdfe83d5fb41ad46c71e762d985b78ee5d061" exitCode=0 Dec 06 04:10:23 crc kubenswrapper[4718]: I1206 04:10:23.763814 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"94e3d13a-7d03-473d-81e7-611590efce2d","Type":"ContainerDied","Data":"8d2bbede898d1fc39452b1b49eefdfe83d5fb41ad46c71e762d985b78ee5d061"} Dec 06 04:10:23 crc kubenswrapper[4718]: I1206 04:10:23.776354 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=14.776340693 podStartE2EDuration="14.776340693s" podCreationTimestamp="2025-12-06 04:10:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:10:23.774666915 +0000 UTC m=+212.780372086" watchObservedRunningTime="2025-12-06 04:10:23.776340693 +0000 UTC m=+212.782045854" Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.068362 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.197968 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/94e3d13a-7d03-473d-81e7-611590efce2d-kubelet-dir\") pod \"94e3d13a-7d03-473d-81e7-611590efce2d\" (UID: \"94e3d13a-7d03-473d-81e7-611590efce2d\") " Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.198092 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94e3d13a-7d03-473d-81e7-611590efce2d-kube-api-access\") pod \"94e3d13a-7d03-473d-81e7-611590efce2d\" (UID: \"94e3d13a-7d03-473d-81e7-611590efce2d\") " Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.198693 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94e3d13a-7d03-473d-81e7-611590efce2d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "94e3d13a-7d03-473d-81e7-611590efce2d" (UID: "94e3d13a-7d03-473d-81e7-611590efce2d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.206658 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e3d13a-7d03-473d-81e7-611590efce2d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "94e3d13a-7d03-473d-81e7-611590efce2d" (UID: "94e3d13a-7d03-473d-81e7-611590efce2d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.300020 4718 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/94e3d13a-7d03-473d-81e7-611590efce2d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.300061 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94e3d13a-7d03-473d-81e7-611590efce2d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.778905 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"94e3d13a-7d03-473d-81e7-611590efce2d","Type":"ContainerDied","Data":"e565c8e716b71b5620f8634a04a797ddfe2c526aa6ec05ea5dd995039c5efc86"} Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.779177 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e565c8e716b71b5620f8634a04a797ddfe2c526aa6ec05ea5dd995039c5efc86" Dec 06 04:10:25 crc kubenswrapper[4718]: I1206 04:10:25.779259 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 04:10:27 crc kubenswrapper[4718]: I1206 04:10:27.877455 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:10:27 crc kubenswrapper[4718]: I1206 04:10:27.878611 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:10:27 crc kubenswrapper[4718]: I1206 04:10:27.878710 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:10:27 crc kubenswrapper[4718]: I1206 04:10:27.879848 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:10:27 crc kubenswrapper[4718]: I1206 04:10:27.882278 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109" gracePeriod=600 Dec 06 04:10:28 crc kubenswrapper[4718]: I1206 04:10:28.797617 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109" exitCode=0 Dec 06 04:10:28 crc kubenswrapper[4718]: I1206 04:10:28.797735 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109"} Dec 06 04:10:28 crc kubenswrapper[4718]: I1206 04:10:28.798373 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"c4e3740d9b890ba2b04c60617f993f3387fea120a80ab270801f84ab026d7dd7"} Dec 06 04:10:28 crc kubenswrapper[4718]: I1206 04:10:28.801015 4718 generic.go:334] "Generic (PLEG): container finished" podID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerID="295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8" exitCode=0 Dec 06 04:10:28 crc kubenswrapper[4718]: I1206 04:10:28.801118 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg6xp" event={"ID":"a8c4e5c8-91b8-432a-bf26-e59627d07754","Type":"ContainerDied","Data":"295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8"} Dec 06 04:10:28 crc kubenswrapper[4718]: I1206 04:10:28.804894 4718 generic.go:334] "Generic (PLEG): container finished" podID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerID="84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7" exitCode=0 Dec 06 04:10:28 crc kubenswrapper[4718]: I1206 04:10:28.804957 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkmtd" event={"ID":"a40acf1b-8c0d-4617-8f41-f48f097cce72","Type":"ContainerDied","Data":"84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7"} Dec 06 04:10:28 crc kubenswrapper[4718]: I1206 04:10:28.818227 4718 generic.go:334] "Generic (PLEG): container finished" podID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerID="c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c" exitCode=0 Dec 06 04:10:28 crc kubenswrapper[4718]: I1206 04:10:28.818329 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdnb" event={"ID":"fbf10d92-c15e-4175-ae97-42e90be68c10","Type":"ContainerDied","Data":"c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c"} Dec 06 04:10:29 crc kubenswrapper[4718]: I1206 04:10:29.824424 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg6xp" event={"ID":"a8c4e5c8-91b8-432a-bf26-e59627d07754","Type":"ContainerStarted","Data":"743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9"} Dec 06 04:10:29 crc kubenswrapper[4718]: I1206 04:10:29.826048 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkmtd" event={"ID":"a40acf1b-8c0d-4617-8f41-f48f097cce72","Type":"ContainerStarted","Data":"569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177"} Dec 06 04:10:29 crc kubenswrapper[4718]: I1206 04:10:29.827853 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdnb" event={"ID":"fbf10d92-c15e-4175-ae97-42e90be68c10","Type":"ContainerStarted","Data":"9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a"} Dec 06 04:10:29 crc kubenswrapper[4718]: I1206 04:10:29.845020 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xg6xp" podStartSLOduration=2.9431221020000002 podStartE2EDuration="1m5.84500552s" podCreationTimestamp="2025-12-06 04:09:24 +0000 UTC" firstStartedPulling="2025-12-06 04:09:26.285595116 +0000 UTC m=+155.291300277" lastFinishedPulling="2025-12-06 04:10:29.187478534 +0000 UTC m=+218.193183695" observedRunningTime="2025-12-06 04:10:29.841241827 +0000 UTC m=+218.846947008" watchObservedRunningTime="2025-12-06 04:10:29.84500552 +0000 UTC m=+218.850710681" Dec 06 04:10:29 crc kubenswrapper[4718]: I1206 04:10:29.861639 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bhdnb" podStartSLOduration=1.94921698 podStartE2EDuration="1m2.861622614s" podCreationTimestamp="2025-12-06 04:09:27 +0000 UTC" firstStartedPulling="2025-12-06 04:09:28.361118401 +0000 UTC m=+157.366823562" lastFinishedPulling="2025-12-06 04:10:29.273524025 +0000 UTC m=+218.279229196" observedRunningTime="2025-12-06 04:10:29.86017606 +0000 UTC m=+218.865881231" watchObservedRunningTime="2025-12-06 04:10:29.861622614 +0000 UTC m=+218.867327775" Dec 06 04:10:32 crc kubenswrapper[4718]: I1206 04:10:32.349201 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nkmtd" podStartSLOduration=5.463156282 podStartE2EDuration="1m6.34918271s" podCreationTimestamp="2025-12-06 04:09:26 +0000 UTC" firstStartedPulling="2025-12-06 04:09:28.314340689 +0000 UTC m=+157.320045850" lastFinishedPulling="2025-12-06 04:10:29.200367117 +0000 UTC m=+218.206072278" observedRunningTime="2025-12-06 04:10:29.878873249 +0000 UTC m=+218.884578420" watchObservedRunningTime="2025-12-06 04:10:32.34918271 +0000 UTC m=+221.354887871" Dec 06 04:10:33 crc kubenswrapper[4718]: I1206 04:10:33.847629 4718 generic.go:334] "Generic (PLEG): container finished" podID="b5790280-687b-4e83-addd-504672823e57" containerID="7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2" exitCode=0 Dec 06 04:10:33 crc kubenswrapper[4718]: I1206 04:10:33.847736 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47mm6" event={"ID":"b5790280-687b-4e83-addd-504672823e57","Type":"ContainerDied","Data":"7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2"} Dec 06 04:10:34 crc kubenswrapper[4718]: I1206 04:10:34.856045 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47mm6" event={"ID":"b5790280-687b-4e83-addd-504672823e57","Type":"ContainerStarted","Data":"370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e"} Dec 06 04:10:34 crc kubenswrapper[4718]: I1206 04:10:34.877171 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-47mm6" podStartSLOduration=2.90311646 podStartE2EDuration="1m7.877155905s" podCreationTimestamp="2025-12-06 04:09:27 +0000 UTC" firstStartedPulling="2025-12-06 04:09:29.368171488 +0000 UTC m=+158.373876649" lastFinishedPulling="2025-12-06 04:10:34.342210923 +0000 UTC m=+223.347916094" observedRunningTime="2025-12-06 04:10:34.875257533 +0000 UTC m=+223.880962694" watchObservedRunningTime="2025-12-06 04:10:34.877155905 +0000 UTC m=+223.882861066" Dec 06 04:10:35 crc kubenswrapper[4718]: I1206 04:10:35.206221 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:10:35 crc kubenswrapper[4718]: I1206 04:10:35.206286 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:10:35 crc kubenswrapper[4718]: I1206 04:10:35.285588 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:10:35 crc kubenswrapper[4718]: I1206 04:10:35.864189 4718 generic.go:334] "Generic (PLEG): container finished" podID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerID="9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049" exitCode=0 Dec 06 04:10:35 crc kubenswrapper[4718]: I1206 04:10:35.864298 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnxxq" event={"ID":"989fbd15-c5fb-41b3-bdc8-2c5399b1424a","Type":"ContainerDied","Data":"9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049"} Dec 06 04:10:35 crc kubenswrapper[4718]: I1206 04:10:35.867502 4718 generic.go:334] "Generic (PLEG): container finished" podID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerID="19d644d2c111dfaa2ddbee73073a4bb9cf0c442562697c846ed4b2b9886c37ff" exitCode=0 Dec 06 04:10:35 crc kubenswrapper[4718]: I1206 04:10:35.867556 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s22px" event={"ID":"360c8eaf-4d22-4e90-acbd-91d78cb1249b","Type":"ContainerDied","Data":"19d644d2c111dfaa2ddbee73073a4bb9cf0c442562697c846ed4b2b9886c37ff"} Dec 06 04:10:35 crc kubenswrapper[4718]: I1206 04:10:35.908674 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:10:36 crc kubenswrapper[4718]: I1206 04:10:36.768474 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xg6xp"] Dec 06 04:10:36 crc kubenswrapper[4718]: I1206 04:10:36.874171 4718 generic.go:334] "Generic (PLEG): container finished" podID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerID="b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2" exitCode=0 Dec 06 04:10:36 crc kubenswrapper[4718]: I1206 04:10:36.874218 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwxl5" event={"ID":"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8","Type":"ContainerDied","Data":"b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2"} Dec 06 04:10:36 crc kubenswrapper[4718]: I1206 04:10:36.878175 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnxxq" event={"ID":"989fbd15-c5fb-41b3-bdc8-2c5399b1424a","Type":"ContainerStarted","Data":"5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0"} Dec 06 04:10:36 crc kubenswrapper[4718]: I1206 04:10:36.886120 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s22px" event={"ID":"360c8eaf-4d22-4e90-acbd-91d78cb1249b","Type":"ContainerStarted","Data":"b6ce372fd21e36f8e7a3cd91aad740c1a9bbe58ec77ab0350167316cea6726f1"} Dec 06 04:10:36 crc kubenswrapper[4718]: I1206 04:10:36.921454 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s22px" podStartSLOduration=1.985416971 podStartE2EDuration="1m8.921434363s" podCreationTimestamp="2025-12-06 04:09:28 +0000 UTC" firstStartedPulling="2025-12-06 04:09:29.351618837 +0000 UTC m=+158.357323998" lastFinishedPulling="2025-12-06 04:10:36.287636239 +0000 UTC m=+225.293341390" observedRunningTime="2025-12-06 04:10:36.919261207 +0000 UTC m=+225.924966368" watchObservedRunningTime="2025-12-06 04:10:36.921434363 +0000 UTC m=+225.927139524" Dec 06 04:10:36 crc kubenswrapper[4718]: I1206 04:10:36.939131 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lnxxq" podStartSLOduration=2.9389987079999997 podStartE2EDuration="1m12.939096154s" podCreationTimestamp="2025-12-06 04:09:24 +0000 UTC" firstStartedPulling="2025-12-06 04:09:26.273202437 +0000 UTC m=+155.278907598" lastFinishedPulling="2025-12-06 04:10:36.273299883 +0000 UTC m=+225.279005044" observedRunningTime="2025-12-06 04:10:36.936131205 +0000 UTC m=+225.941836386" watchObservedRunningTime="2025-12-06 04:10:36.939096154 +0000 UTC m=+225.944801315" Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.008312 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.009224 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.048431 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.438407 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.439005 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.478994 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.567825 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb297"] Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.896801 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwxl5" event={"ID":"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8","Type":"ContainerStarted","Data":"1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353"} Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.897170 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xg6xp" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerName="registry-server" containerID="cri-o://743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9" gracePeriod=2 Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.918214 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwxl5" podStartSLOduration=2.938478439 podStartE2EDuration="1m13.918199099s" podCreationTimestamp="2025-12-06 04:09:24 +0000 UTC" firstStartedPulling="2025-12-06 04:09:26.283310282 +0000 UTC m=+155.289015443" lastFinishedPulling="2025-12-06 04:10:37.263030942 +0000 UTC m=+226.268736103" observedRunningTime="2025-12-06 04:10:37.917290033 +0000 UTC m=+226.922995194" watchObservedRunningTime="2025-12-06 04:10:37.918199099 +0000 UTC m=+226.923904250" Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.953598 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:10:37 crc kubenswrapper[4718]: I1206 04:10:37.953648 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.058653 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.059241 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.328963 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.436685 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.436727 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.469172 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs9mk\" (UniqueName: \"kubernetes.io/projected/a8c4e5c8-91b8-432a-bf26-e59627d07754-kube-api-access-zs9mk\") pod \"a8c4e5c8-91b8-432a-bf26-e59627d07754\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.469263 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-catalog-content\") pod \"a8c4e5c8-91b8-432a-bf26-e59627d07754\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.469359 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-utilities\") pod \"a8c4e5c8-91b8-432a-bf26-e59627d07754\" (UID: \"a8c4e5c8-91b8-432a-bf26-e59627d07754\") " Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.470152 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-utilities" (OuterVolumeSpecName: "utilities") pod "a8c4e5c8-91b8-432a-bf26-e59627d07754" (UID: "a8c4e5c8-91b8-432a-bf26-e59627d07754"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.495500 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8c4e5c8-91b8-432a-bf26-e59627d07754-kube-api-access-zs9mk" (OuterVolumeSpecName: "kube-api-access-zs9mk") pod "a8c4e5c8-91b8-432a-bf26-e59627d07754" (UID: "a8c4e5c8-91b8-432a-bf26-e59627d07754"). InnerVolumeSpecName "kube-api-access-zs9mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.531918 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8c4e5c8-91b8-432a-bf26-e59627d07754" (UID: "a8c4e5c8-91b8-432a-bf26-e59627d07754"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.570476 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs9mk\" (UniqueName: \"kubernetes.io/projected/a8c4e5c8-91b8-432a-bf26-e59627d07754-kube-api-access-zs9mk\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.570520 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.570537 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c4e5c8-91b8-432a-bf26-e59627d07754-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.913780 4718 generic.go:334] "Generic (PLEG): container finished" podID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerID="b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18" exitCode=0 Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.913844 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pxg" event={"ID":"3bbaf3bc-767a-4bfc-ad55-42c198b0b990","Type":"ContainerDied","Data":"b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18"} Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.917886 4718 generic.go:334] "Generic (PLEG): container finished" podID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerID="743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9" exitCode=0 Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.917923 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg6xp" event={"ID":"a8c4e5c8-91b8-432a-bf26-e59627d07754","Type":"ContainerDied","Data":"743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9"} Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.917966 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xg6xp" event={"ID":"a8c4e5c8-91b8-432a-bf26-e59627d07754","Type":"ContainerDied","Data":"249db4e214fdbfcdce2aaf059a05ed45704c219d3831210bb855761c61ca9931"} Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.917978 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xg6xp" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.917988 4718 scope.go:117] "RemoveContainer" containerID="743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.946778 4718 scope.go:117] "RemoveContainer" containerID="295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8" Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.957910 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xg6xp"] Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.960564 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xg6xp"] Dec 06 04:10:38 crc kubenswrapper[4718]: I1206 04:10:38.980903 4718 scope.go:117] "RemoveContainer" containerID="660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531" Dec 06 04:10:39 crc kubenswrapper[4718]: I1206 04:10:39.008157 4718 scope.go:117] "RemoveContainer" containerID="743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9" Dec 06 04:10:39 crc kubenswrapper[4718]: E1206 04:10:39.008825 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9\": container with ID starting with 743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9 not found: ID does not exist" containerID="743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9" Dec 06 04:10:39 crc kubenswrapper[4718]: I1206 04:10:39.008886 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9"} err="failed to get container status \"743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9\": rpc error: code = NotFound desc = could not find container \"743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9\": container with ID starting with 743daadbb518a6249d83c28bb88a0d295b86ba450412d91e34a4c1fcf08ecdf9 not found: ID does not exist" Dec 06 04:10:39 crc kubenswrapper[4718]: I1206 04:10:39.008960 4718 scope.go:117] "RemoveContainer" containerID="295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8" Dec 06 04:10:39 crc kubenswrapper[4718]: E1206 04:10:39.009362 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8\": container with ID starting with 295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8 not found: ID does not exist" containerID="295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8" Dec 06 04:10:39 crc kubenswrapper[4718]: I1206 04:10:39.009404 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8"} err="failed to get container status \"295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8\": rpc error: code = NotFound desc = could not find container \"295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8\": container with ID starting with 295eb797e8a1b0f436ff3d7cbb79069c97d86519ae7a061f8e84dbed3d3354a8 not found: ID does not exist" Dec 06 04:10:39 crc kubenswrapper[4718]: I1206 04:10:39.009432 4718 scope.go:117] "RemoveContainer" containerID="660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531" Dec 06 04:10:39 crc kubenswrapper[4718]: E1206 04:10:39.009736 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531\": container with ID starting with 660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531 not found: ID does not exist" containerID="660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531" Dec 06 04:10:39 crc kubenswrapper[4718]: I1206 04:10:39.009769 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531"} err="failed to get container status \"660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531\": rpc error: code = NotFound desc = could not find container \"660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531\": container with ID starting with 660d64c8079feeaf0e3247e4361c9685487889020d3a04e0d99ae6b7e29f8531 not found: ID does not exist" Dec 06 04:10:39 crc kubenswrapper[4718]: I1206 04:10:39.102220 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-47mm6" podUID="b5790280-687b-4e83-addd-504672823e57" containerName="registry-server" probeResult="failure" output=< Dec 06 04:10:39 crc kubenswrapper[4718]: timeout: failed to connect service ":50051" within 1s Dec 06 04:10:39 crc kubenswrapper[4718]: > Dec 06 04:10:39 crc kubenswrapper[4718]: I1206 04:10:39.336661 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" path="/var/lib/kubelet/pods/a8c4e5c8-91b8-432a-bf26-e59627d07754/volumes" Dec 06 04:10:39 crc kubenswrapper[4718]: I1206 04:10:39.478506 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s22px" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerName="registry-server" probeResult="failure" output=< Dec 06 04:10:39 crc kubenswrapper[4718]: timeout: failed to connect service ":50051" within 1s Dec 06 04:10:39 crc kubenswrapper[4718]: > Dec 06 04:10:41 crc kubenswrapper[4718]: I1206 04:10:41.369197 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdnb"] Dec 06 04:10:41 crc kubenswrapper[4718]: I1206 04:10:41.369771 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bhdnb" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerName="registry-server" containerID="cri-o://9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a" gracePeriod=2 Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.836143 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.926948 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-utilities\") pod \"fbf10d92-c15e-4175-ae97-42e90be68c10\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.927020 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjvdt\" (UniqueName: \"kubernetes.io/projected/fbf10d92-c15e-4175-ae97-42e90be68c10-kube-api-access-fjvdt\") pod \"fbf10d92-c15e-4175-ae97-42e90be68c10\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.927081 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-catalog-content\") pod \"fbf10d92-c15e-4175-ae97-42e90be68c10\" (UID: \"fbf10d92-c15e-4175-ae97-42e90be68c10\") " Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.927806 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-utilities" (OuterVolumeSpecName: "utilities") pod "fbf10d92-c15e-4175-ae97-42e90be68c10" (UID: "fbf10d92-c15e-4175-ae97-42e90be68c10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.939710 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbf10d92-c15e-4175-ae97-42e90be68c10-kube-api-access-fjvdt" (OuterVolumeSpecName: "kube-api-access-fjvdt") pod "fbf10d92-c15e-4175-ae97-42e90be68c10" (UID: "fbf10d92-c15e-4175-ae97-42e90be68c10"). InnerVolumeSpecName "kube-api-access-fjvdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.945737 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fbf10d92-c15e-4175-ae97-42e90be68c10" (UID: "fbf10d92-c15e-4175-ae97-42e90be68c10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.945953 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pxg" event={"ID":"3bbaf3bc-767a-4bfc-ad55-42c198b0b990","Type":"ContainerStarted","Data":"536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4"} Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.949340 4718 generic.go:334] "Generic (PLEG): container finished" podID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerID="9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a" exitCode=0 Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.949375 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdnb" event={"ID":"fbf10d92-c15e-4175-ae97-42e90be68c10","Type":"ContainerDied","Data":"9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a"} Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.949401 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdnb" event={"ID":"fbf10d92-c15e-4175-ae97-42e90be68c10","Type":"ContainerDied","Data":"f4ec6c7ae75bad2987a9212f021079783050b2c520bb4a990a0296f415b541a9"} Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.949417 4718 scope.go:117] "RemoveContainer" containerID="9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a" Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.949481 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhdnb" Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.963596 4718 scope.go:117] "RemoveContainer" containerID="c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c" Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.974485 4718 scope.go:117] "RemoveContainer" containerID="43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971" Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.975582 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c4pxg" podStartSLOduration=2.048367001 podStartE2EDuration="1m17.97556875s" podCreationTimestamp="2025-12-06 04:09:25 +0000 UTC" firstStartedPulling="2025-12-06 04:09:26.280003576 +0000 UTC m=+155.285708737" lastFinishedPulling="2025-12-06 04:10:42.207205325 +0000 UTC m=+231.212910486" observedRunningTime="2025-12-06 04:10:42.964464956 +0000 UTC m=+231.970170117" watchObservedRunningTime="2025-12-06 04:10:42.97556875 +0000 UTC m=+231.981273911" Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.987879 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdnb"] Dec 06 04:10:42 crc kubenswrapper[4718]: I1206 04:10:42.993655 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdnb"] Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.008742 4718 scope.go:117] "RemoveContainer" containerID="9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a" Dec 06 04:10:43 crc kubenswrapper[4718]: E1206 04:10:43.009276 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a\": container with ID starting with 9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a not found: ID does not exist" containerID="9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a" Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.009355 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a"} err="failed to get container status \"9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a\": rpc error: code = NotFound desc = could not find container \"9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a\": container with ID starting with 9e34fea30c0166886cadedb76acc93568ceaaa762a2a2952d17e5efc20fcd62a not found: ID does not exist" Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.009396 4718 scope.go:117] "RemoveContainer" containerID="c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c" Dec 06 04:10:43 crc kubenswrapper[4718]: E1206 04:10:43.009915 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c\": container with ID starting with c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c not found: ID does not exist" containerID="c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c" Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.009962 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c"} err="failed to get container status \"c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c\": rpc error: code = NotFound desc = could not find container \"c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c\": container with ID starting with c0c180b96956933562779205ef8cdf003e9805adc6341d4dc66bbee4a807ec4c not found: ID does not exist" Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.009997 4718 scope.go:117] "RemoveContainer" containerID="43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971" Dec 06 04:10:43 crc kubenswrapper[4718]: E1206 04:10:43.010310 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971\": container with ID starting with 43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971 not found: ID does not exist" containerID="43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971" Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.010346 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971"} err="failed to get container status \"43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971\": rpc error: code = NotFound desc = could not find container \"43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971\": container with ID starting with 43aa8806ea2ffecdd8067e0a17472d3f80519c375b1311a38265536470bbe971 not found: ID does not exist" Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.029494 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.029543 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjvdt\" (UniqueName: \"kubernetes.io/projected/fbf10d92-c15e-4175-ae97-42e90be68c10-kube-api-access-fjvdt\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.029555 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf10d92-c15e-4175-ae97-42e90be68c10-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:43 crc kubenswrapper[4718]: I1206 04:10:43.334829 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" path="/var/lib/kubelet/pods/fbf10d92-c15e-4175-ae97-42e90be68c10/volumes" Dec 06 04:10:44 crc kubenswrapper[4718]: I1206 04:10:44.798638 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:10:44 crc kubenswrapper[4718]: I1206 04:10:44.799645 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:10:44 crc kubenswrapper[4718]: I1206 04:10:44.848868 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:10:44 crc kubenswrapper[4718]: I1206 04:10:44.998850 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:10:45 crc kubenswrapper[4718]: I1206 04:10:45.003610 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:10:45 crc kubenswrapper[4718]: I1206 04:10:45.003728 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:10:45 crc kubenswrapper[4718]: I1206 04:10:45.049410 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:10:45 crc kubenswrapper[4718]: I1206 04:10:45.407962 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:10:45 crc kubenswrapper[4718]: I1206 04:10:45.408264 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:10:45 crc kubenswrapper[4718]: I1206 04:10:45.464911 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:10:46 crc kubenswrapper[4718]: I1206 04:10:46.027999 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:10:48 crc kubenswrapper[4718]: I1206 04:10:48.106549 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:10:48 crc kubenswrapper[4718]: I1206 04:10:48.163892 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:10:48 crc kubenswrapper[4718]: I1206 04:10:48.479641 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:10:48 crc kubenswrapper[4718]: I1206 04:10:48.527988 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:10:50 crc kubenswrapper[4718]: I1206 04:10:50.387495 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s22px"] Dec 06 04:10:50 crc kubenswrapper[4718]: I1206 04:10:50.388284 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s22px" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerName="registry-server" containerID="cri-o://b6ce372fd21e36f8e7a3cd91aad740c1a9bbe58ec77ab0350167316cea6726f1" gracePeriod=2 Dec 06 04:10:53 crc kubenswrapper[4718]: I1206 04:10:53.014696 4718 generic.go:334] "Generic (PLEG): container finished" podID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerID="b6ce372fd21e36f8e7a3cd91aad740c1a9bbe58ec77ab0350167316cea6726f1" exitCode=0 Dec 06 04:10:53 crc kubenswrapper[4718]: I1206 04:10:53.014760 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s22px" event={"ID":"360c8eaf-4d22-4e90-acbd-91d78cb1249b","Type":"ContainerDied","Data":"b6ce372fd21e36f8e7a3cd91aad740c1a9bbe58ec77ab0350167316cea6726f1"} Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.459047 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.519889 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c4pxg"] Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.637785 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.703347 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpx9f\" (UniqueName: \"kubernetes.io/projected/360c8eaf-4d22-4e90-acbd-91d78cb1249b-kube-api-access-vpx9f\") pod \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.703464 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-utilities\") pod \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.703521 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-catalog-content\") pod \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\" (UID: \"360c8eaf-4d22-4e90-acbd-91d78cb1249b\") " Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.704819 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-utilities" (OuterVolumeSpecName: "utilities") pod "360c8eaf-4d22-4e90-acbd-91d78cb1249b" (UID: "360c8eaf-4d22-4e90-acbd-91d78cb1249b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.713306 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/360c8eaf-4d22-4e90-acbd-91d78cb1249b-kube-api-access-vpx9f" (OuterVolumeSpecName: "kube-api-access-vpx9f") pod "360c8eaf-4d22-4e90-acbd-91d78cb1249b" (UID: "360c8eaf-4d22-4e90-acbd-91d78cb1249b"). InnerVolumeSpecName "kube-api-access-vpx9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.805471 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpx9f\" (UniqueName: \"kubernetes.io/projected/360c8eaf-4d22-4e90-acbd-91d78cb1249b-kube-api-access-vpx9f\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.805524 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.811003 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "360c8eaf-4d22-4e90-acbd-91d78cb1249b" (UID: "360c8eaf-4d22-4e90-acbd-91d78cb1249b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:10:55 crc kubenswrapper[4718]: I1206 04:10:55.906723 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/360c8eaf-4d22-4e90-acbd-91d78cb1249b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.054109 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s22px" event={"ID":"360c8eaf-4d22-4e90-acbd-91d78cb1249b","Type":"ContainerDied","Data":"7c4fd01697de5a4e7dd940dc6917a69b7eb586724d8f15a1772e750255d33308"} Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.054161 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s22px" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.054198 4718 scope.go:117] "RemoveContainer" containerID="b6ce372fd21e36f8e7a3cd91aad740c1a9bbe58ec77ab0350167316cea6726f1" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.054388 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c4pxg" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerName="registry-server" containerID="cri-o://536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4" gracePeriod=2 Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.079822 4718 scope.go:117] "RemoveContainer" containerID="19d644d2c111dfaa2ddbee73073a4bb9cf0c442562697c846ed4b2b9886c37ff" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.103744 4718 scope.go:117] "RemoveContainer" containerID="7a429892afc695df0cae3397cf4770f0591564d67de90a8ab8d8e988c955484d" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.113529 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s22px"] Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.118646 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s22px"] Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.719977 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.818632 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6bvs\" (UniqueName: \"kubernetes.io/projected/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-kube-api-access-w6bvs\") pod \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.818691 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-catalog-content\") pod \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.818766 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-utilities\") pod \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\" (UID: \"3bbaf3bc-767a-4bfc-ad55-42c198b0b990\") " Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.819698 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-utilities" (OuterVolumeSpecName: "utilities") pod "3bbaf3bc-767a-4bfc-ad55-42c198b0b990" (UID: "3bbaf3bc-767a-4bfc-ad55-42c198b0b990"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.823083 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-kube-api-access-w6bvs" (OuterVolumeSpecName: "kube-api-access-w6bvs") pod "3bbaf3bc-767a-4bfc-ad55-42c198b0b990" (UID: "3bbaf3bc-767a-4bfc-ad55-42c198b0b990"). InnerVolumeSpecName "kube-api-access-w6bvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.880161 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bbaf3bc-767a-4bfc-ad55-42c198b0b990" (UID: "3bbaf3bc-767a-4bfc-ad55-42c198b0b990"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.920084 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.920140 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6bvs\" (UniqueName: \"kubernetes.io/projected/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-kube-api-access-w6bvs\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:56 crc kubenswrapper[4718]: I1206 04:10:56.920163 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bbaf3bc-767a-4bfc-ad55-42c198b0b990-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.065648 4718 generic.go:334] "Generic (PLEG): container finished" podID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerID="536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4" exitCode=0 Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.065705 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pxg" event={"ID":"3bbaf3bc-767a-4bfc-ad55-42c198b0b990","Type":"ContainerDied","Data":"536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4"} Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.065732 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c4pxg" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.065754 4718 scope.go:117] "RemoveContainer" containerID="536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.065741 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c4pxg" event={"ID":"3bbaf3bc-767a-4bfc-ad55-42c198b0b990","Type":"ContainerDied","Data":"d07cb4f8beafef3e7cebcc3ca940feb51f69a2c032cef6044fb23263a23d6b81"} Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.088103 4718 scope.go:117] "RemoveContainer" containerID="b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.096514 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c4pxg"] Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.099915 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c4pxg"] Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.119147 4718 scope.go:117] "RemoveContainer" containerID="0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.136607 4718 scope.go:117] "RemoveContainer" containerID="536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4" Dec 06 04:10:57 crc kubenswrapper[4718]: E1206 04:10:57.137220 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4\": container with ID starting with 536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4 not found: ID does not exist" containerID="536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.137268 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4"} err="failed to get container status \"536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4\": rpc error: code = NotFound desc = could not find container \"536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4\": container with ID starting with 536b761937c5f70130ed75eacf5e53ba4895878793ce6ea4e2399294a85045f4 not found: ID does not exist" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.137292 4718 scope.go:117] "RemoveContainer" containerID="b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18" Dec 06 04:10:57 crc kubenswrapper[4718]: E1206 04:10:57.138334 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18\": container with ID starting with b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18 not found: ID does not exist" containerID="b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.138390 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18"} err="failed to get container status \"b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18\": rpc error: code = NotFound desc = could not find container \"b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18\": container with ID starting with b7c1fcce2406da0c0f33ec0ca4339c7f9e31b47312e091f09096b7dd44ebfb18 not found: ID does not exist" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.138426 4718 scope.go:117] "RemoveContainer" containerID="0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f" Dec 06 04:10:57 crc kubenswrapper[4718]: E1206 04:10:57.138856 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f\": container with ID starting with 0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f not found: ID does not exist" containerID="0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.138883 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f"} err="failed to get container status \"0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f\": rpc error: code = NotFound desc = could not find container \"0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f\": container with ID starting with 0bc10e393a5eb562c0b64fa6bbaf9ff8413027f274bc0e8e4cf044624cad2a3f not found: ID does not exist" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.336982 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" path="/var/lib/kubelet/pods/360c8eaf-4d22-4e90-acbd-91d78cb1249b/volumes" Dec 06 04:10:57 crc kubenswrapper[4718]: I1206 04:10:57.337776 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" path="/var/lib/kubelet/pods/3bbaf3bc-767a-4bfc-ad55-42c198b0b990/volumes" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.716296 4718 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717008 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerName="extract-content" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717030 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerName="extract-content" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717048 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerName="extract-utilities" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717060 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerName="extract-utilities" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717079 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerName="extract-utilities" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717094 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerName="extract-utilities" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717110 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717121 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717142 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717152 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717168 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e3d13a-7d03-473d-81e7-611590efce2d" containerName="pruner" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717180 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e3d13a-7d03-473d-81e7-611590efce2d" containerName="pruner" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717193 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerName="extract-utilities" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717204 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerName="extract-utilities" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717216 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerName="extract-content" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717247 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerName="extract-content" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717269 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717280 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717299 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerName="extract-content" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717312 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerName="extract-content" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717330 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerName="extract-utilities" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717342 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerName="extract-utilities" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717357 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerName="extract-content" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717369 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerName="extract-content" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.717382 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717393 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717565 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bbaf3bc-767a-4bfc-ad55-42c198b0b990" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717588 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8c4e5c8-91b8-432a-bf26-e59627d07754" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717606 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="360c8eaf-4d22-4e90-acbd-91d78cb1249b" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717627 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e3d13a-7d03-473d-81e7-611590efce2d" containerName="pruner" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.717644 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf10d92-c15e-4175-ae97-42e90be68c10" containerName="registry-server" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.718158 4718 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.718366 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.718609 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb" gracePeriod=15 Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.718707 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2" gracePeriod=15 Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.718707 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9" gracePeriod=15 Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.718888 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e" gracePeriod=15 Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.718955 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6" gracePeriod=15 Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722058 4718 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.722399 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722432 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.722448 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722460 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.722474 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722485 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.722508 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722519 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.722539 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722550 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.722565 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722576 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722789 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722819 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722838 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722854 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.722871 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 04:11:00 crc kubenswrapper[4718]: E1206 04:11:00.723072 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.723087 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.723347 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.770589 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.771087 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.771177 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.771214 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.771342 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.771432 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.771464 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.771522 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.844691 4718 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.844754 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872412 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872457 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872479 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872500 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872534 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872551 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872586 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872639 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872705 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872748 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872775 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872801 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872825 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872852 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872878 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:00 crc kubenswrapper[4718]: I1206 04:11:00.872906 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.100072 4718 generic.go:334] "Generic (PLEG): container finished" podID="bde102c1-6414-4a3a-b463-8f57a44257c7" containerID="c60f02bc09933f7fee870bb3350a4550280b59696c6dbd814755940d1cf395e6" exitCode=0 Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.100135 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bde102c1-6414-4a3a-b463-8f57a44257c7","Type":"ContainerDied","Data":"c60f02bc09933f7fee870bb3350a4550280b59696c6dbd814755940d1cf395e6"} Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.101329 4718 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.101750 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.103272 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.104976 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.105806 4718 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e" exitCode=0 Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.105842 4718 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6" exitCode=0 Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.105859 4718 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2" exitCode=0 Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.105873 4718 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9" exitCode=2 Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.105940 4718 scope.go:117] "RemoveContainer" containerID="f89a801e0a2e42a6250ce9608387a42a64d52cb894a2fb9558c15190675d09f8" Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.331960 4718 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:01 crc kubenswrapper[4718]: I1206 04:11:01.332445 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.120133 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.415362 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.416300 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.503550 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-var-lock\") pod \"bde102c1-6414-4a3a-b463-8f57a44257c7\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.503618 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bde102c1-6414-4a3a-b463-8f57a44257c7-kube-api-access\") pod \"bde102c1-6414-4a3a-b463-8f57a44257c7\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.503662 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-var-lock" (OuterVolumeSpecName: "var-lock") pod "bde102c1-6414-4a3a-b463-8f57a44257c7" (UID: "bde102c1-6414-4a3a-b463-8f57a44257c7"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.503713 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-kubelet-dir\") pod \"bde102c1-6414-4a3a-b463-8f57a44257c7\" (UID: \"bde102c1-6414-4a3a-b463-8f57a44257c7\") " Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.503875 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bde102c1-6414-4a3a-b463-8f57a44257c7" (UID: "bde102c1-6414-4a3a-b463-8f57a44257c7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.504097 4718 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.504127 4718 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bde102c1-6414-4a3a-b463-8f57a44257c7-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.511818 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde102c1-6414-4a3a-b463-8f57a44257c7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bde102c1-6414-4a3a-b463-8f57a44257c7" (UID: "bde102c1-6414-4a3a-b463-8f57a44257c7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.592197 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" containerName="oauth-openshift" containerID="cri-o://0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde" gracePeriod=15 Dec 06 04:11:02 crc kubenswrapper[4718]: I1206 04:11:02.605752 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bde102c1-6414-4a3a-b463-8f57a44257c7-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.106783 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.107706 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.108313 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.110984 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.111904 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.112617 4718 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.113044 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.113344 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.127979 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.128029 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bde102c1-6414-4a3a-b463-8f57a44257c7","Type":"ContainerDied","Data":"2874a166a5b920c5e1a33551548b16c0bae0e5eb166b6f4d1e464878574ab026"} Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.128068 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2874a166a5b920c5e1a33551548b16c0bae0e5eb166b6f4d1e464878574ab026" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.135678 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.136654 4718 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb" exitCode=0 Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.136752 4718 scope.go:117] "RemoveContainer" containerID="c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.136772 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.138627 4718 generic.go:334] "Generic (PLEG): container finished" podID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" containerID="0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde" exitCode=0 Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.138671 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" event={"ID":"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a","Type":"ContainerDied","Data":"0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde"} Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.138691 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.138701 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" event={"ID":"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a","Type":"ContainerDied","Data":"95eaac06549b409c9d323ab1f97442e4314a6c77799b463c9feeefdff41bb7b5"} Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.140379 4718 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.140693 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.141075 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.142202 4718 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.142497 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.142720 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.151532 4718 scope.go:117] "RemoveContainer" containerID="205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.161185 4718 scope.go:117] "RemoveContainer" containerID="41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.171804 4718 scope.go:117] "RemoveContainer" containerID="7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.182926 4718 scope.go:117] "RemoveContainer" containerID="3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.197141 4718 scope.go:117] "RemoveContainer" containerID="d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.213055 4718 scope.go:117] "RemoveContainer" containerID="c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e" Dec 06 04:11:03 crc kubenswrapper[4718]: E1206 04:11:03.213542 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\": container with ID starting with c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e not found: ID does not exist" containerID="c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.213602 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e"} err="failed to get container status \"c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\": rpc error: code = NotFound desc = could not find container \"c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e\": container with ID starting with c8db18af59821dd289205d154a635d33e1e3bf891b57209ad4abd95d4f1bf88e not found: ID does not exist" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.213633 4718 scope.go:117] "RemoveContainer" containerID="205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6" Dec 06 04:11:03 crc kubenswrapper[4718]: E1206 04:11:03.213995 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\": container with ID starting with 205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6 not found: ID does not exist" containerID="205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.214052 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6"} err="failed to get container status \"205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\": rpc error: code = NotFound desc = could not find container \"205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6\": container with ID starting with 205841adc1c0b9d06d4d4cbecc6ef1a0617dc981d108b55397f66411fcd088e6 not found: ID does not exist" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.214076 4718 scope.go:117] "RemoveContainer" containerID="41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2" Dec 06 04:11:03 crc kubenswrapper[4718]: E1206 04:11:03.214388 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\": container with ID starting with 41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2 not found: ID does not exist" containerID="41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.214408 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2"} err="failed to get container status \"41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\": rpc error: code = NotFound desc = could not find container \"41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2\": container with ID starting with 41f90377c8a766a9da383024c196b1047ae0c723a7f9873dc5aed3add08fa4e2 not found: ID does not exist" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.214421 4718 scope.go:117] "RemoveContainer" containerID="7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9" Dec 06 04:11:03 crc kubenswrapper[4718]: E1206 04:11:03.214632 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\": container with ID starting with 7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9 not found: ID does not exist" containerID="7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.214654 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9"} err="failed to get container status \"7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\": rpc error: code = NotFound desc = could not find container \"7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9\": container with ID starting with 7376e8f1a7a44b3f631372261521b8000fbef3d5afd4a103ab7a39f3bb471fd9 not found: ID does not exist" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.214668 4718 scope.go:117] "RemoveContainer" containerID="3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb" Dec 06 04:11:03 crc kubenswrapper[4718]: E1206 04:11:03.214965 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\": container with ID starting with 3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb not found: ID does not exist" containerID="3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215007 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb"} err="failed to get container status \"3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\": rpc error: code = NotFound desc = could not find container \"3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb\": container with ID starting with 3f0a6efef523542cede36bbe1ae20f507ae55633603832a71febca64638e7bfb not found: ID does not exist" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215036 4718 scope.go:117] "RemoveContainer" containerID="d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215154 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-router-certs\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215184 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-cliconfig\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215205 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-idp-0-file-data\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215275 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-error\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215299 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-ocp-branding-template\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215314 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215334 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-trusted-ca-bundle\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215359 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-serving-cert\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215375 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-service-ca\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215388 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-dir\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215401 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215417 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215441 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-provider-selection\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215461 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-login\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215479 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqmfw\" (UniqueName: \"kubernetes.io/projected/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-kube-api-access-kqmfw\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215499 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-session\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215517 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-policies\") pod \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\" (UID: \"3f24f572-933e-4ccc-a9d6-ea8cf40aee2a\") " Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215549 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215922 4718 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215964 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.215975 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.216092 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.216128 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: E1206 04:11:03.215447 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\": container with ID starting with d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13 not found: ID does not exist" containerID="d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.216144 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.216159 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.216336 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13"} err="failed to get container status \"d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\": rpc error: code = NotFound desc = could not find container \"d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13\": container with ID starting with d51f8bf1d6f467179da76943ac9b287d85e76c2f34a2eeda939cbd506841eb13 not found: ID does not exist" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.216361 4718 scope.go:117] "RemoveContainer" containerID="0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.216681 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.219310 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.219528 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.219783 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.219899 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.220037 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-kube-api-access-kqmfw" (OuterVolumeSpecName: "kube-api-access-kqmfw") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "kube-api-access-kqmfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.220442 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.220514 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.220682 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.221719 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" (UID: "3f24f572-933e-4ccc-a9d6-ea8cf40aee2a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.234305 4718 scope.go:117] "RemoveContainer" containerID="0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde" Dec 06 04:11:03 crc kubenswrapper[4718]: E1206 04:11:03.234688 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde\": container with ID starting with 0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde not found: ID does not exist" containerID="0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.234713 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde"} err="failed to get container status \"0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde\": rpc error: code = NotFound desc = could not find container \"0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde\": container with ID starting with 0beae14d50bbddd231da7488f060788257978f9001b330a54e4f7756797a2fde not found: ID does not exist" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317113 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317161 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317176 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqmfw\" (UniqueName: \"kubernetes.io/projected/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-kube-api-access-kqmfw\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317185 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317197 4718 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317206 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317217 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317229 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317255 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317264 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317280 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317294 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317306 4718 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317316 4718 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317325 4718 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.317332 4718 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.336430 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.441575 4718 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.442253 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.442789 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.443889 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.444449 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:03 crc kubenswrapper[4718]: I1206 04:11:03.444828 4718 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:05 crc kubenswrapper[4718]: E1206 04:11:05.764581 4718 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:05 crc kubenswrapper[4718]: I1206 04:11:05.765667 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:05 crc kubenswrapper[4718]: E1206 04:11:05.804989 4718 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e84eef3d298fd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 04:11:05.803561213 +0000 UTC m=+254.809266424,LastTimestamp:2025-12-06 04:11:05.803561213 +0000 UTC m=+254.809266424,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 04:11:06 crc kubenswrapper[4718]: I1206 04:11:06.165063 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b91129b191dda1f023c87514e217f1e01e48daf05f02e87d30f1d7e3067d2223"} Dec 06 04:11:07 crc kubenswrapper[4718]: I1206 04:11:07.174973 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7"} Dec 06 04:11:07 crc kubenswrapper[4718]: E1206 04:11:07.175871 4718 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:07 crc kubenswrapper[4718]: I1206 04:11:07.175854 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:07 crc kubenswrapper[4718]: I1206 04:11:07.176611 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:07 crc kubenswrapper[4718]: E1206 04:11:07.255252 4718 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:07 crc kubenswrapper[4718]: E1206 04:11:07.255730 4718 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:07 crc kubenswrapper[4718]: E1206 04:11:07.255997 4718 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:07 crc kubenswrapper[4718]: E1206 04:11:07.256445 4718 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:07 crc kubenswrapper[4718]: E1206 04:11:07.256689 4718 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:07 crc kubenswrapper[4718]: I1206 04:11:07.256719 4718 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 06 04:11:07 crc kubenswrapper[4718]: E1206 04:11:07.256922 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="200ms" Dec 06 04:11:07 crc kubenswrapper[4718]: E1206 04:11:07.457779 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="400ms" Dec 06 04:11:07 crc kubenswrapper[4718]: E1206 04:11:07.859138 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="800ms" Dec 06 04:11:08 crc kubenswrapper[4718]: E1206 04:11:08.182421 4718 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:09 crc kubenswrapper[4718]: E1206 04:11:09.217619 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="1.6s" Dec 06 04:11:10 crc kubenswrapper[4718]: E1206 04:11:10.818738 4718 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="3.2s" Dec 06 04:11:11 crc kubenswrapper[4718]: I1206 04:11:11.331008 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:11 crc kubenswrapper[4718]: I1206 04:11:11.331441 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:12 crc kubenswrapper[4718]: I1206 04:11:12.327914 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:12 crc kubenswrapper[4718]: I1206 04:11:12.328902 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:12 crc kubenswrapper[4718]: I1206 04:11:12.329520 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:12 crc kubenswrapper[4718]: I1206 04:11:12.345822 4718 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:12 crc kubenswrapper[4718]: I1206 04:11:12.346104 4718 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:12 crc kubenswrapper[4718]: E1206 04:11:12.346562 4718 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:12 crc kubenswrapper[4718]: I1206 04:11:12.347099 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:12 crc kubenswrapper[4718]: E1206 04:11:12.600048 4718 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e84eef3d298fd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 04:11:05.803561213 +0000 UTC m=+254.809266424,LastTimestamp:2025-12-06 04:11:05.803561213 +0000 UTC m=+254.809266424,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 04:11:13 crc kubenswrapper[4718]: I1206 04:11:13.272902 4718 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="6d732858c771d42bf877d831949a60f66c43f758136d64b174f2e4920bed6c17" exitCode=0 Dec 06 04:11:13 crc kubenswrapper[4718]: I1206 04:11:13.273032 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"6d732858c771d42bf877d831949a60f66c43f758136d64b174f2e4920bed6c17"} Dec 06 04:11:13 crc kubenswrapper[4718]: I1206 04:11:13.273820 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6a9f79dd85e7e48d6cbb64df6441b5c26e816a423a083679de2aa8df281ec952"} Dec 06 04:11:13 crc kubenswrapper[4718]: I1206 04:11:13.274420 4718 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:13 crc kubenswrapper[4718]: I1206 04:11:13.274466 4718 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:13 crc kubenswrapper[4718]: E1206 04:11:13.275033 4718 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:13 crc kubenswrapper[4718]: I1206 04:11:13.275457 4718 status_manager.go:851] "Failed to get status for pod" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" pod="openshift-authentication/oauth-openshift-558db77b4-pb297" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pb297\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:13 crc kubenswrapper[4718]: I1206 04:11:13.276105 4718 status_manager.go:851] "Failed to get status for pod" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Dec 06 04:11:14 crc kubenswrapper[4718]: I1206 04:11:14.284077 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 04:11:14 crc kubenswrapper[4718]: I1206 04:11:14.284442 4718 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17" exitCode=1 Dec 06 04:11:14 crc kubenswrapper[4718]: I1206 04:11:14.284500 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17"} Dec 06 04:11:14 crc kubenswrapper[4718]: I1206 04:11:14.285004 4718 scope.go:117] "RemoveContainer" containerID="d039dc43ce0c9aebf6c8d5db6acff849518823c25fcfd3a258ca18b5aefcdf17" Dec 06 04:11:14 crc kubenswrapper[4718]: I1206 04:11:14.295737 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"30a8ef60d7ba98b36ffccb4598905e1df85e2e9a956bdb480564773a65485ba4"} Dec 06 04:11:14 crc kubenswrapper[4718]: I1206 04:11:14.295780 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3c1bb94a0ccb058948974f182a478105bca2783854ae43ed70aa51a1f550a9f3"} Dec 06 04:11:14 crc kubenswrapper[4718]: I1206 04:11:14.295790 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b2fa902e55acc09010573f84529bf10b1abaadce093b04d3b77acb18198bfbbf"} Dec 06 04:11:15 crc kubenswrapper[4718]: I1206 04:11:15.309950 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dcc28d1e472bd2ae1eb422eecd5272167cbf604a88ebcbe469cc4eb72b3688b6"} Dec 06 04:11:15 crc kubenswrapper[4718]: I1206 04:11:15.309991 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"06173a4aec3fd0e9b46f92abf1bc1d02b227ea93245b494a2ae22a448de25564"} Dec 06 04:11:15 crc kubenswrapper[4718]: I1206 04:11:15.310148 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:15 crc kubenswrapper[4718]: I1206 04:11:15.310339 4718 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:15 crc kubenswrapper[4718]: I1206 04:11:15.310375 4718 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:15 crc kubenswrapper[4718]: I1206 04:11:15.313984 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 04:11:15 crc kubenswrapper[4718]: I1206 04:11:15.314083 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8febe3477a84bbae3b445af33205e8c4d8b9a8c111f9a5086916933c9cf95a5d"} Dec 06 04:11:16 crc kubenswrapper[4718]: I1206 04:11:16.838416 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:11:16 crc kubenswrapper[4718]: I1206 04:11:16.845104 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:11:17 crc kubenswrapper[4718]: I1206 04:11:17.338569 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:11:17 crc kubenswrapper[4718]: I1206 04:11:17.347289 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:17 crc kubenswrapper[4718]: I1206 04:11:17.347341 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:17 crc kubenswrapper[4718]: I1206 04:11:17.359880 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:20 crc kubenswrapper[4718]: I1206 04:11:20.322952 4718 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:21 crc kubenswrapper[4718]: I1206 04:11:21.347404 4718 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:21 crc kubenswrapper[4718]: I1206 04:11:21.347702 4718 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:21 crc kubenswrapper[4718]: I1206 04:11:21.353338 4718 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a1b1d0c3-c2ce-4427-b2fe-075e7c735183" Dec 06 04:11:21 crc kubenswrapper[4718]: I1206 04:11:21.353581 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:22 crc kubenswrapper[4718]: I1206 04:11:22.353545 4718 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:22 crc kubenswrapper[4718]: I1206 04:11:22.354932 4718 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:22 crc kubenswrapper[4718]: I1206 04:11:22.356455 4718 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a1b1d0c3-c2ce-4427-b2fe-075e7c735183" Dec 06 04:11:29 crc kubenswrapper[4718]: I1206 04:11:29.534665 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 04:11:30 crc kubenswrapper[4718]: I1206 04:11:30.169194 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 04:11:30 crc kubenswrapper[4718]: I1206 04:11:30.751852 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 04:11:31 crc kubenswrapper[4718]: I1206 04:11:31.047453 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 04:11:31 crc kubenswrapper[4718]: I1206 04:11:31.194266 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 04:11:31 crc kubenswrapper[4718]: I1206 04:11:31.566887 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 04:11:31 crc kubenswrapper[4718]: I1206 04:11:31.809982 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 04:11:32 crc kubenswrapper[4718]: I1206 04:11:32.220708 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 04:11:32 crc kubenswrapper[4718]: I1206 04:11:32.542694 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 04:11:32 crc kubenswrapper[4718]: I1206 04:11:32.694346 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 04:11:32 crc kubenswrapper[4718]: I1206 04:11:32.850752 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 04:11:32 crc kubenswrapper[4718]: I1206 04:11:32.946146 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.260875 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.280891 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.450027 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.450268 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.605116 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.642466 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.665791 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.793056 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.836973 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.961812 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 04:11:33 crc kubenswrapper[4718]: I1206 04:11:33.965328 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.097912 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.112276 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.272462 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.343266 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.381847 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.423628 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.506128 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.540927 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.565780 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.669783 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 04:11:34 crc kubenswrapper[4718]: I1206 04:11:34.844675 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.032007 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.174943 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.260437 4718 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.304078 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.334613 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.374065 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.417055 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.450033 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.502828 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.511125 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.703291 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.718108 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.852297 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.973043 4718 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 04:11:35 crc kubenswrapper[4718]: I1206 04:11:35.982259 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.007620 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.044639 4718 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.195701 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.309741 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.402712 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.539877 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.650179 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.653388 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.783735 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.941725 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 04:11:36 crc kubenswrapper[4718]: I1206 04:11:36.956706 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.031139 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.064588 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.256612 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.275298 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.339160 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.354483 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.355837 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.385202 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.540485 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.576136 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.608544 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.635888 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.695335 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.721788 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.778981 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.781654 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.782164 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 04:11:37 crc kubenswrapper[4718]: I1206 04:11:37.980832 4718 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.012664 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.146195 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.166495 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.271110 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.287539 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.334527 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.381166 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.433341 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.440456 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.465725 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.553753 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.565816 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.624399 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.797104 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.842215 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.880788 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.953227 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.970178 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 04:11:38 crc kubenswrapper[4718]: I1206 04:11:38.991419 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.007963 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.026904 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.040122 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.057545 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.218053 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.233890 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.267503 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.375539 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.423025 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.520435 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.739458 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.746974 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.785405 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.873862 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.925051 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.952783 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.954353 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 04:11:39 crc kubenswrapper[4718]: I1206 04:11:39.955261 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.040122 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.046637 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.148354 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.301762 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.440280 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.500761 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.504442 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.512433 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.517919 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.589622 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.608380 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.622485 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.655618 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.686473 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.717758 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.726055 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.747797 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.821352 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.940262 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 04:11:40 crc kubenswrapper[4718]: I1206 04:11:40.952620 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.074921 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.178316 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.191344 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.224225 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.270490 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.384590 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.505695 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.574345 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.604580 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.640188 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.663638 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.845010 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.905659 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 04:11:41 crc kubenswrapper[4718]: I1206 04:11:41.920397 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.008445 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.040788 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.061737 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.073930 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.111856 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.112122 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.148496 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.180342 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.198495 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.256113 4718 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.262282 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb297","openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.262362 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l","openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 04:11:42 crc kubenswrapper[4718]: E1206 04:11:42.262603 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" containerName="oauth-openshift" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.262626 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" containerName="oauth-openshift" Dec 06 04:11:42 crc kubenswrapper[4718]: E1206 04:11:42.262648 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" containerName="installer" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.262657 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" containerName="installer" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.262792 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde102c1-6414-4a3a-b463-8f57a44257c7" containerName="installer" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.262810 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" containerName="oauth-openshift" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.263088 4718 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.263154 4718 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="245fe818-e203-4799-93f1-1f96b6972225" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.263377 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.268389 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.268842 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.269193 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.269734 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.270160 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.270427 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.271027 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.271342 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.271682 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.271760 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.272360 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.273767 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.273870 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.275728 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.280985 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.284751 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.295135 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.300879 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.300853512 podStartE2EDuration="22.300853512s" podCreationTimestamp="2025-12-06 04:11:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:11:42.300797901 +0000 UTC m=+291.306503102" watchObservedRunningTime="2025-12-06 04:11:42.300853512 +0000 UTC m=+291.306558683" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.304882 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-service-ca\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.304975 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305045 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305076 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305137 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305168 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdbzx\" (UniqueName: \"kubernetes.io/projected/10c8fa31-5b14-4cd5-8703-304c79ef4897-kube-api-access-sdbzx\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305262 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-session\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305389 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-router-certs\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305462 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305535 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-template-error\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305614 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-audit-policies\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305718 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-template-login\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305773 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10c8fa31-5b14-4cd5-8703-304c79ef4897-audit-dir\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.305807 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.310567 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407052 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-template-error\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407111 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-audit-policies\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407173 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-template-login\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407202 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10c8fa31-5b14-4cd5-8703-304c79ef4897-audit-dir\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407223 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407279 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-service-ca\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407312 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407347 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407368 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407397 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10c8fa31-5b14-4cd5-8703-304c79ef4897-audit-dir\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.407410 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.408071 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdbzx\" (UniqueName: \"kubernetes.io/projected/10c8fa31-5b14-4cd5-8703-304c79ef4897-kube-api-access-sdbzx\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.408118 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-session\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.408163 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-router-certs\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.408209 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.408279 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-service-ca\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.408403 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.409356 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-audit-policies\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.409388 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.412164 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.412185 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-template-error\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.412584 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-template-login\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.412959 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-router-certs\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.412993 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.413913 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.414895 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.416475 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/10c8fa31-5b14-4cd5-8703-304c79ef4897-v4-0-config-system-session\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.422475 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.430133 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdbzx\" (UniqueName: \"kubernetes.io/projected/10c8fa31-5b14-4cd5-8703-304c79ef4897-kube-api-access-sdbzx\") pod \"oauth-openshift-b6fcd9dcb-qxh5l\" (UID: \"10c8fa31-5b14-4cd5-8703-304c79ef4897\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.542182 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.602326 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.604753 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.706955 4718 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.707459 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7" gracePeriod=5 Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.725781 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.796902 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.841620 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l"] Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.886572 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.912412 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.919436 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.956819 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.983662 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 04:11:42 crc kubenswrapper[4718]: I1206 04:11:42.987153 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.038704 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.040255 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.056456 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.091118 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.129375 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.164408 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.182775 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.273969 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.335711 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f24f572-933e-4ccc-a9d6-ea8cf40aee2a" path="/var/lib/kubelet/pods/3f24f572-933e-4ccc-a9d6-ea8cf40aee2a/volumes" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.345738 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.388775 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.480173 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" event={"ID":"10c8fa31-5b14-4cd5-8703-304c79ef4897","Type":"ContainerStarted","Data":"9181313453a2bf4da76fdcfa9f9fb4eac23c8d850f054377be6ca12f03528aff"} Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.481695 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" event={"ID":"10c8fa31-5b14-4cd5-8703-304c79ef4897","Type":"ContainerStarted","Data":"c250922cc7ba0c1668f3c786b2e2290b99ad790570cf3833820e56a2f95f9af3"} Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.504135 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" podStartSLOduration=66.504115453 podStartE2EDuration="1m6.504115453s" podCreationTimestamp="2025-12-06 04:10:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:11:43.503859198 +0000 UTC m=+292.509564419" watchObservedRunningTime="2025-12-06 04:11:43.504115453 +0000 UTC m=+292.509820604" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.549209 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.711292 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.750903 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.869998 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.942492 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 04:11:43 crc kubenswrapper[4718]: I1206 04:11:43.960259 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.032446 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.039389 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.087652 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.117832 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.149326 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.149894 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.397579 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.406124 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.487604 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.493587 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-qxh5l" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.671178 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.803516 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 04:11:44 crc kubenswrapper[4718]: I1206 04:11:44.922699 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.006126 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.023261 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.069983 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.081875 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.106750 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.208148 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.429570 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.516842 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.529837 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.700160 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 04:11:45 crc kubenswrapper[4718]: I1206 04:11:45.816163 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.015720 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.088470 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.093700 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.344303 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.385709 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.472533 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.486143 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.538814 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.541448 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.556594 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.594852 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.767735 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 04:11:46 crc kubenswrapper[4718]: I1206 04:11:46.937131 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 04:11:47 crc kubenswrapper[4718]: I1206 04:11:47.045081 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 04:11:47 crc kubenswrapper[4718]: I1206 04:11:47.079953 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 04:11:47 crc kubenswrapper[4718]: I1206 04:11:47.557257 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 04:11:47 crc kubenswrapper[4718]: I1206 04:11:47.841123 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 04:11:47 crc kubenswrapper[4718]: I1206 04:11:47.928147 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 04:11:47 crc kubenswrapper[4718]: I1206 04:11:47.964220 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.039852 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.102989 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.236358 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.313882 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.313962 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.383743 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.383884 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.383894 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.383990 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.384205 4718 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.384222 4718 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.432503 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.481940 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.484835 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.484880 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.484903 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.485132 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.485203 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.493446 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.511980 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.512067 4718 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7" exitCode=137 Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.512156 4718 scope.go:117] "RemoveContainer" containerID="20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.512387 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.543345 4718 scope.go:117] "RemoveContainer" containerID="20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7" Dec 06 04:11:48 crc kubenswrapper[4718]: E1206 04:11:48.543864 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7\": container with ID starting with 20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7 not found: ID does not exist" containerID="20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.543895 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7"} err="failed to get container status \"20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7\": rpc error: code = NotFound desc = could not find container \"20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7\": container with ID starting with 20e0435b209e029ab777b7e08c287f90617af7aa9f430274ba7e87caa1c702d7 not found: ID does not exist" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.586706 4718 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.586745 4718 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:48 crc kubenswrapper[4718]: I1206 04:11:48.586755 4718 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 06 04:11:49 crc kubenswrapper[4718]: I1206 04:11:49.270950 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 04:11:49 crc kubenswrapper[4718]: I1206 04:11:49.338427 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 06 04:11:49 crc kubenswrapper[4718]: I1206 04:11:49.505714 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 04:11:49 crc kubenswrapper[4718]: I1206 04:11:49.776110 4718 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 04:11:50 crc kubenswrapper[4718]: I1206 04:11:50.169516 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 04:12:23 crc kubenswrapper[4718]: I1206 04:12:23.452012 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7fbd"] Dec 06 04:12:23 crc kubenswrapper[4718]: I1206 04:12:23.452800 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" podUID="b13c7e13-af0d-4c28-bb48-e457a608075e" containerName="controller-manager" containerID="cri-o://04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08" gracePeriod=30 Dec 06 04:12:23 crc kubenswrapper[4718]: I1206 04:12:23.573317 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx"] Dec 06 04:12:23 crc kubenswrapper[4718]: I1206 04:12:23.573568 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" podUID="c76e70f2-707c-4d55-8af2-672fac1f7ac4" containerName="route-controller-manager" containerID="cri-o://6ffed2f61374089653bb845fb2e38803382615ca3ea948ed4b24a4b221627811" gracePeriod=30 Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.470959 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.607450 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-proxy-ca-bundles\") pod \"b13c7e13-af0d-4c28-bb48-e457a608075e\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.607508 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b13c7e13-af0d-4c28-bb48-e457a608075e-serving-cert\") pod \"b13c7e13-af0d-4c28-bb48-e457a608075e\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.607577 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-config\") pod \"b13c7e13-af0d-4c28-bb48-e457a608075e\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.607600 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8m87\" (UniqueName: \"kubernetes.io/projected/b13c7e13-af0d-4c28-bb48-e457a608075e-kube-api-access-l8m87\") pod \"b13c7e13-af0d-4c28-bb48-e457a608075e\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.607640 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-client-ca\") pod \"b13c7e13-af0d-4c28-bb48-e457a608075e\" (UID: \"b13c7e13-af0d-4c28-bb48-e457a608075e\") " Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.608710 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b13c7e13-af0d-4c28-bb48-e457a608075e" (UID: "b13c7e13-af0d-4c28-bb48-e457a608075e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.609192 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-client-ca" (OuterVolumeSpecName: "client-ca") pod "b13c7e13-af0d-4c28-bb48-e457a608075e" (UID: "b13c7e13-af0d-4c28-bb48-e457a608075e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.609640 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-config" (OuterVolumeSpecName: "config") pod "b13c7e13-af0d-4c28-bb48-e457a608075e" (UID: "b13c7e13-af0d-4c28-bb48-e457a608075e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.616989 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b13c7e13-af0d-4c28-bb48-e457a608075e-kube-api-access-l8m87" (OuterVolumeSpecName: "kube-api-access-l8m87") pod "b13c7e13-af0d-4c28-bb48-e457a608075e" (UID: "b13c7e13-af0d-4c28-bb48-e457a608075e"). InnerVolumeSpecName "kube-api-access-l8m87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.619427 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b13c7e13-af0d-4c28-bb48-e457a608075e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b13c7e13-af0d-4c28-bb48-e457a608075e" (UID: "b13c7e13-af0d-4c28-bb48-e457a608075e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.709138 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.709183 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8m87\" (UniqueName: \"kubernetes.io/projected/b13c7e13-af0d-4c28-bb48-e457a608075e-kube-api-access-l8m87\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.709193 4718 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.709203 4718 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b13c7e13-af0d-4c28-bb48-e457a608075e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.709214 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b13c7e13-af0d-4c28-bb48-e457a608075e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.737361 4718 generic.go:334] "Generic (PLEG): container finished" podID="b13c7e13-af0d-4c28-bb48-e457a608075e" containerID="04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08" exitCode=0 Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.737417 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.737437 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" event={"ID":"b13c7e13-af0d-4c28-bb48-e457a608075e","Type":"ContainerDied","Data":"04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08"} Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.737466 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v7fbd" event={"ID":"b13c7e13-af0d-4c28-bb48-e457a608075e","Type":"ContainerDied","Data":"d61de62e8aa3c22f6fa2ab09211b6b1e1e6033532f37fbb5fec14c346bd0386c"} Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.737482 4718 scope.go:117] "RemoveContainer" containerID="04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.740169 4718 generic.go:334] "Generic (PLEG): container finished" podID="c76e70f2-707c-4d55-8af2-672fac1f7ac4" containerID="6ffed2f61374089653bb845fb2e38803382615ca3ea948ed4b24a4b221627811" exitCode=0 Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.740193 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" event={"ID":"c76e70f2-707c-4d55-8af2-672fac1f7ac4","Type":"ContainerDied","Data":"6ffed2f61374089653bb845fb2e38803382615ca3ea948ed4b24a4b221627811"} Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.755350 4718 scope.go:117] "RemoveContainer" containerID="04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08" Dec 06 04:12:24 crc kubenswrapper[4718]: E1206 04:12:24.756211 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08\": container with ID starting with 04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08 not found: ID does not exist" containerID="04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.756342 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08"} err="failed to get container status \"04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08\": rpc error: code = NotFound desc = could not find container \"04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08\": container with ID starting with 04c1cc44c85b8ecdc2cd08450fda1a27386d6cec2c6ba3cf200f48ab14c6ac08 not found: ID does not exist" Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.765833 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7fbd"] Dec 06 04:12:24 crc kubenswrapper[4718]: I1206 04:12:24.769153 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7fbd"] Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.333714 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b13c7e13-af0d-4c28-bb48-e457a608075e" path="/var/lib/kubelet/pods/b13c7e13-af0d-4c28-bb48-e457a608075e/volumes" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.514405 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb"] Dec 06 04:12:25 crc kubenswrapper[4718]: E1206 04:12:25.514647 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.514662 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 04:12:25 crc kubenswrapper[4718]: E1206 04:12:25.514689 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13c7e13-af0d-4c28-bb48-e457a608075e" containerName="controller-manager" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.514697 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13c7e13-af0d-4c28-bb48-e457a608075e" containerName="controller-manager" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.514801 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.514816 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="b13c7e13-af0d-4c28-bb48-e457a608075e" containerName="controller-manager" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.515273 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.518582 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.520164 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.521140 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.521161 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.523721 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.524325 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.532763 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb"] Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.544657 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.617983 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fxsn\" (UniqueName: \"kubernetes.io/projected/4bbfde19-028f-469a-8af1-e2b28334b256-kube-api-access-2fxsn\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.618044 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbfde19-028f-469a-8af1-e2b28334b256-serving-cert\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.618067 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-proxy-ca-bundles\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.618093 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-config\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.618125 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-client-ca\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.719059 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fxsn\" (UniqueName: \"kubernetes.io/projected/4bbfde19-028f-469a-8af1-e2b28334b256-kube-api-access-2fxsn\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.719136 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbfde19-028f-469a-8af1-e2b28334b256-serving-cert\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.719175 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-proxy-ca-bundles\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.719222 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-config\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.719319 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-client-ca\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.720561 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-proxy-ca-bundles\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.720716 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-client-ca\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.721053 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-config\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.724718 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbfde19-028f-469a-8af1-e2b28334b256-serving-cert\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.745284 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fxsn\" (UniqueName: \"kubernetes.io/projected/4bbfde19-028f-469a-8af1-e2b28334b256-kube-api-access-2fxsn\") pod \"controller-manager-6cdffcbc8f-m28tb\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:25 crc kubenswrapper[4718]: I1206 04:12:25.834999 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:26 crc kubenswrapper[4718]: I1206 04:12:26.158873 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb"] Dec 06 04:12:26 crc kubenswrapper[4718]: I1206 04:12:26.755704 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" event={"ID":"4bbfde19-028f-469a-8af1-e2b28334b256","Type":"ContainerStarted","Data":"c619e5f9f64c323846edd776857bfd89b3cf6646c3647e06f0605e81af076c85"} Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.111102 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.241547 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c76e70f2-707c-4d55-8af2-672fac1f7ac4-serving-cert\") pod \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.241609 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-config\") pod \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.241646 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlgqt\" (UniqueName: \"kubernetes.io/projected/c76e70f2-707c-4d55-8af2-672fac1f7ac4-kube-api-access-mlgqt\") pod \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.241670 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-client-ca\") pod \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\" (UID: \"c76e70f2-707c-4d55-8af2-672fac1f7ac4\") " Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.242400 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-client-ca" (OuterVolumeSpecName: "client-ca") pod "c76e70f2-707c-4d55-8af2-672fac1f7ac4" (UID: "c76e70f2-707c-4d55-8af2-672fac1f7ac4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.242423 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-config" (OuterVolumeSpecName: "config") pod "c76e70f2-707c-4d55-8af2-672fac1f7ac4" (UID: "c76e70f2-707c-4d55-8af2-672fac1f7ac4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.247736 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c76e70f2-707c-4d55-8af2-672fac1f7ac4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c76e70f2-707c-4d55-8af2-672fac1f7ac4" (UID: "c76e70f2-707c-4d55-8af2-672fac1f7ac4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.258652 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c76e70f2-707c-4d55-8af2-672fac1f7ac4-kube-api-access-mlgqt" (OuterVolumeSpecName: "kube-api-access-mlgqt") pod "c76e70f2-707c-4d55-8af2-672fac1f7ac4" (UID: "c76e70f2-707c-4d55-8af2-672fac1f7ac4"). InnerVolumeSpecName "kube-api-access-mlgqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.343047 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c76e70f2-707c-4d55-8af2-672fac1f7ac4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.343086 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.343094 4718 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c76e70f2-707c-4d55-8af2-672fac1f7ac4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.343104 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlgqt\" (UniqueName: \"kubernetes.io/projected/c76e70f2-707c-4d55-8af2-672fac1f7ac4-kube-api-access-mlgqt\") on node \"crc\" DevicePath \"\"" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.512392 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp"] Dec 06 04:12:27 crc kubenswrapper[4718]: E1206 04:12:27.512725 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c76e70f2-707c-4d55-8af2-672fac1f7ac4" containerName="route-controller-manager" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.512747 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="c76e70f2-707c-4d55-8af2-672fac1f7ac4" containerName="route-controller-manager" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.512898 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="c76e70f2-707c-4d55-8af2-672fac1f7ac4" containerName="route-controller-manager" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.513545 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.519005 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp"] Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.646801 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlb25\" (UniqueName: \"kubernetes.io/projected/8e598e76-0b15-4420-b806-1157a59f8e59-kube-api-access-jlb25\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.646854 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e598e76-0b15-4420-b806-1157a59f8e59-config\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.646890 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e598e76-0b15-4420-b806-1157a59f8e59-serving-cert\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.646937 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e598e76-0b15-4420-b806-1157a59f8e59-client-ca\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.747646 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e598e76-0b15-4420-b806-1157a59f8e59-serving-cert\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.747750 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e598e76-0b15-4420-b806-1157a59f8e59-client-ca\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.747788 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlb25\" (UniqueName: \"kubernetes.io/projected/8e598e76-0b15-4420-b806-1157a59f8e59-kube-api-access-jlb25\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.747820 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e598e76-0b15-4420-b806-1157a59f8e59-config\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.748956 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e598e76-0b15-4420-b806-1157a59f8e59-client-ca\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.749390 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e598e76-0b15-4420-b806-1157a59f8e59-config\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.752202 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e598e76-0b15-4420-b806-1157a59f8e59-serving-cert\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.763787 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" event={"ID":"4bbfde19-028f-469a-8af1-e2b28334b256","Type":"ContainerStarted","Data":"4b8b3314ee24706e3c876756b2c1141324b8252e6d840d0804d87008c0c9189e"} Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.765448 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.766839 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" event={"ID":"c76e70f2-707c-4d55-8af2-672fac1f7ac4","Type":"ContainerDied","Data":"1696c6b496fd16d0f3365033ce45862d2f40359388f21ded797563687cb94687"} Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.766896 4718 scope.go:117] "RemoveContainer" containerID="6ffed2f61374089653bb845fb2e38803382615ca3ea948ed4b24a4b221627811" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.766985 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.775825 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlb25\" (UniqueName: \"kubernetes.io/projected/8e598e76-0b15-4420-b806-1157a59f8e59-kube-api-access-jlb25\") pod \"route-controller-manager-7c4c5c5554-d4qvp\" (UID: \"8e598e76-0b15-4420-b806-1157a59f8e59\") " pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.778573 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.793891 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" podStartSLOduration=4.793872835 podStartE2EDuration="4.793872835s" podCreationTimestamp="2025-12-06 04:12:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:12:27.790803794 +0000 UTC m=+336.796509005" watchObservedRunningTime="2025-12-06 04:12:27.793872835 +0000 UTC m=+336.799578006" Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.823076 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx"] Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.828079 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfkx"] Dec 06 04:12:27 crc kubenswrapper[4718]: I1206 04:12:27.828405 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:28 crc kubenswrapper[4718]: I1206 04:12:28.020633 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp"] Dec 06 04:12:28 crc kubenswrapper[4718]: I1206 04:12:28.773654 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" event={"ID":"8e598e76-0b15-4420-b806-1157a59f8e59","Type":"ContainerStarted","Data":"68747b2a828e6981f6e74f4df5a82b516b8af4a3da96576dbf72b75e57738094"} Dec 06 04:12:28 crc kubenswrapper[4718]: I1206 04:12:28.773928 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" event={"ID":"8e598e76-0b15-4420-b806-1157a59f8e59","Type":"ContainerStarted","Data":"f0307c1cb89694493b32477021c190fe172dc594d3776b6d2eafad475081ef79"} Dec 06 04:12:28 crc kubenswrapper[4718]: I1206 04:12:28.791738 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" podStartSLOduration=5.791718405 podStartE2EDuration="5.791718405s" podCreationTimestamp="2025-12-06 04:12:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:12:28.787495673 +0000 UTC m=+337.793200844" watchObservedRunningTime="2025-12-06 04:12:28.791718405 +0000 UTC m=+337.797423566" Dec 06 04:12:29 crc kubenswrapper[4718]: I1206 04:12:29.337525 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c76e70f2-707c-4d55-8af2-672fac1f7ac4" path="/var/lib/kubelet/pods/c76e70f2-707c-4d55-8af2-672fac1f7ac4/volumes" Dec 06 04:12:29 crc kubenswrapper[4718]: I1206 04:12:29.779126 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:29 crc kubenswrapper[4718]: I1206 04:12:29.787368 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7c4c5c5554-d4qvp" Dec 06 04:12:57 crc kubenswrapper[4718]: I1206 04:12:57.877390 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:12:57 crc kubenswrapper[4718]: I1206 04:12:57.878081 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.223496 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fbpfb"] Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.224606 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.242259 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fbpfb"] Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.385352 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/646f2bb7-2bdb-430d-b08b-b11f49bdb356-trusted-ca\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.385398 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/646f2bb7-2bdb-430d-b08b-b11f49bdb356-registry-certificates\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.385421 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/646f2bb7-2bdb-430d-b08b-b11f49bdb356-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.385469 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.385496 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/646f2bb7-2bdb-430d-b08b-b11f49bdb356-registry-tls\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.385675 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/646f2bb7-2bdb-430d-b08b-b11f49bdb356-bound-sa-token\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.385792 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crlbc\" (UniqueName: \"kubernetes.io/projected/646f2bb7-2bdb-430d-b08b-b11f49bdb356-kube-api-access-crlbc\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.385854 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/646f2bb7-2bdb-430d-b08b-b11f49bdb356-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.416901 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.486526 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/646f2bb7-2bdb-430d-b08b-b11f49bdb356-trusted-ca\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.486570 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/646f2bb7-2bdb-430d-b08b-b11f49bdb356-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.486588 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/646f2bb7-2bdb-430d-b08b-b11f49bdb356-registry-certificates\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.486634 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/646f2bb7-2bdb-430d-b08b-b11f49bdb356-registry-tls\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.486681 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/646f2bb7-2bdb-430d-b08b-b11f49bdb356-bound-sa-token\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.486717 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crlbc\" (UniqueName: \"kubernetes.io/projected/646f2bb7-2bdb-430d-b08b-b11f49bdb356-kube-api-access-crlbc\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.486749 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/646f2bb7-2bdb-430d-b08b-b11f49bdb356-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.488184 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/646f2bb7-2bdb-430d-b08b-b11f49bdb356-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.488462 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/646f2bb7-2bdb-430d-b08b-b11f49bdb356-trusted-ca\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.489037 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/646f2bb7-2bdb-430d-b08b-b11f49bdb356-registry-certificates\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.492006 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/646f2bb7-2bdb-430d-b08b-b11f49bdb356-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.501716 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/646f2bb7-2bdb-430d-b08b-b11f49bdb356-bound-sa-token\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.502179 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/646f2bb7-2bdb-430d-b08b-b11f49bdb356-registry-tls\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.506540 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crlbc\" (UniqueName: \"kubernetes.io/projected/646f2bb7-2bdb-430d-b08b-b11f49bdb356-kube-api-access-crlbc\") pod \"image-registry-66df7c8f76-fbpfb\" (UID: \"646f2bb7-2bdb-430d-b08b-b11f49bdb356\") " pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.550321 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:13 crc kubenswrapper[4718]: I1206 04:13:13.972257 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fbpfb"] Dec 06 04:13:13 crc kubenswrapper[4718]: W1206 04:13:13.978869 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod646f2bb7_2bdb_430d_b08b_b11f49bdb356.slice/crio-d855749fc89d8a7582772eda349222734aa483197d4bb80e2a2a17e6fa226dcb WatchSource:0}: Error finding container d855749fc89d8a7582772eda349222734aa483197d4bb80e2a2a17e6fa226dcb: Status 404 returned error can't find the container with id d855749fc89d8a7582772eda349222734aa483197d4bb80e2a2a17e6fa226dcb Dec 06 04:13:14 crc kubenswrapper[4718]: I1206 04:13:14.073858 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" event={"ID":"646f2bb7-2bdb-430d-b08b-b11f49bdb356","Type":"ContainerStarted","Data":"d855749fc89d8a7582772eda349222734aa483197d4bb80e2a2a17e6fa226dcb"} Dec 06 04:13:15 crc kubenswrapper[4718]: I1206 04:13:15.080137 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" event={"ID":"646f2bb7-2bdb-430d-b08b-b11f49bdb356","Type":"ContainerStarted","Data":"5a55a64cae7d742cfdfc1c39fadf3fd778466b3e12b534c1bdd10f9d0758c3c7"} Dec 06 04:13:15 crc kubenswrapper[4718]: I1206 04:13:15.080475 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:15 crc kubenswrapper[4718]: I1206 04:13:15.099115 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" podStartSLOduration=2.099100239 podStartE2EDuration="2.099100239s" podCreationTimestamp="2025-12-06 04:13:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:13:15.095818213 +0000 UTC m=+384.101523374" watchObservedRunningTime="2025-12-06 04:13:15.099100239 +0000 UTC m=+384.104805400" Dec 06 04:13:23 crc kubenswrapper[4718]: I1206 04:13:23.462653 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb"] Dec 06 04:13:23 crc kubenswrapper[4718]: I1206 04:13:23.464109 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" podUID="4bbfde19-028f-469a-8af1-e2b28334b256" containerName="controller-manager" containerID="cri-o://4b8b3314ee24706e3c876756b2c1141324b8252e6d840d0804d87008c0c9189e" gracePeriod=30 Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.136185 4718 generic.go:334] "Generic (PLEG): container finished" podID="4bbfde19-028f-469a-8af1-e2b28334b256" containerID="4b8b3314ee24706e3c876756b2c1141324b8252e6d840d0804d87008c0c9189e" exitCode=0 Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.136259 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" event={"ID":"4bbfde19-028f-469a-8af1-e2b28334b256","Type":"ContainerDied","Data":"4b8b3314ee24706e3c876756b2c1141324b8252e6d840d0804d87008c0c9189e"} Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.319411 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.443022 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbfde19-028f-469a-8af1-e2b28334b256-serving-cert\") pod \"4bbfde19-028f-469a-8af1-e2b28334b256\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.443077 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-proxy-ca-bundles\") pod \"4bbfde19-028f-469a-8af1-e2b28334b256\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.443166 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-config\") pod \"4bbfde19-028f-469a-8af1-e2b28334b256\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.443279 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-client-ca\") pod \"4bbfde19-028f-469a-8af1-e2b28334b256\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.443327 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fxsn\" (UniqueName: \"kubernetes.io/projected/4bbfde19-028f-469a-8af1-e2b28334b256-kube-api-access-2fxsn\") pod \"4bbfde19-028f-469a-8af1-e2b28334b256\" (UID: \"4bbfde19-028f-469a-8af1-e2b28334b256\") " Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.444074 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4bbfde19-028f-469a-8af1-e2b28334b256" (UID: "4bbfde19-028f-469a-8af1-e2b28334b256"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.444213 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-config" (OuterVolumeSpecName: "config") pod "4bbfde19-028f-469a-8af1-e2b28334b256" (UID: "4bbfde19-028f-469a-8af1-e2b28334b256"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.444206 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-client-ca" (OuterVolumeSpecName: "client-ca") pod "4bbfde19-028f-469a-8af1-e2b28334b256" (UID: "4bbfde19-028f-469a-8af1-e2b28334b256"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.451484 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbfde19-028f-469a-8af1-e2b28334b256-kube-api-access-2fxsn" (OuterVolumeSpecName: "kube-api-access-2fxsn") pod "4bbfde19-028f-469a-8af1-e2b28334b256" (UID: "4bbfde19-028f-469a-8af1-e2b28334b256"). InnerVolumeSpecName "kube-api-access-2fxsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.451549 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbfde19-028f-469a-8af1-e2b28334b256-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4bbfde19-028f-469a-8af1-e2b28334b256" (UID: "4bbfde19-028f-469a-8af1-e2b28334b256"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.544256 4718 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbfde19-028f-469a-8af1-e2b28334b256-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.544291 4718 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.544301 4718 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.544309 4718 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbfde19-028f-469a-8af1-e2b28334b256-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.544318 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fxsn\" (UniqueName: \"kubernetes.io/projected/4bbfde19-028f-469a-8af1-e2b28334b256-kube-api-access-2fxsn\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.544256 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b7b55f4df-j9qjj"] Dec 06 04:13:24 crc kubenswrapper[4718]: E1206 04:13:24.544494 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbfde19-028f-469a-8af1-e2b28334b256" containerName="controller-manager" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.544504 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbfde19-028f-469a-8af1-e2b28334b256" containerName="controller-manager" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.544600 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbfde19-028f-469a-8af1-e2b28334b256" containerName="controller-manager" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.544961 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.561789 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b7b55f4df-j9qjj"] Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.645551 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/704e278f-04a5-431b-b716-e3a2e71afc10-client-ca\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.645638 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/704e278f-04a5-431b-b716-e3a2e71afc10-serving-cert\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.645674 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/704e278f-04a5-431b-b716-e3a2e71afc10-config\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.645703 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttc2w\" (UniqueName: \"kubernetes.io/projected/704e278f-04a5-431b-b716-e3a2e71afc10-kube-api-access-ttc2w\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.645765 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/704e278f-04a5-431b-b716-e3a2e71afc10-proxy-ca-bundles\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.747435 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/704e278f-04a5-431b-b716-e3a2e71afc10-proxy-ca-bundles\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.747532 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/704e278f-04a5-431b-b716-e3a2e71afc10-client-ca\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.747592 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/704e278f-04a5-431b-b716-e3a2e71afc10-serving-cert\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.747613 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/704e278f-04a5-431b-b716-e3a2e71afc10-config\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.747650 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttc2w\" (UniqueName: \"kubernetes.io/projected/704e278f-04a5-431b-b716-e3a2e71afc10-kube-api-access-ttc2w\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.748546 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/704e278f-04a5-431b-b716-e3a2e71afc10-client-ca\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.748918 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/704e278f-04a5-431b-b716-e3a2e71afc10-proxy-ca-bundles\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.749781 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/704e278f-04a5-431b-b716-e3a2e71afc10-config\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.752268 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/704e278f-04a5-431b-b716-e3a2e71afc10-serving-cert\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.763884 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttc2w\" (UniqueName: \"kubernetes.io/projected/704e278f-04a5-431b-b716-e3a2e71afc10-kube-api-access-ttc2w\") pod \"controller-manager-b7b55f4df-j9qjj\" (UID: \"704e278f-04a5-431b-b716-e3a2e71afc10\") " pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:24 crc kubenswrapper[4718]: I1206 04:13:24.873992 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:25 crc kubenswrapper[4718]: I1206 04:13:25.064938 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b7b55f4df-j9qjj"] Dec 06 04:13:25 crc kubenswrapper[4718]: I1206 04:13:25.141860 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" Dec 06 04:13:25 crc kubenswrapper[4718]: I1206 04:13:25.141904 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb" event={"ID":"4bbfde19-028f-469a-8af1-e2b28334b256","Type":"ContainerDied","Data":"c619e5f9f64c323846edd776857bfd89b3cf6646c3647e06f0605e81af076c85"} Dec 06 04:13:25 crc kubenswrapper[4718]: I1206 04:13:25.141988 4718 scope.go:117] "RemoveContainer" containerID="4b8b3314ee24706e3c876756b2c1141324b8252e6d840d0804d87008c0c9189e" Dec 06 04:13:25 crc kubenswrapper[4718]: I1206 04:13:25.144120 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" event={"ID":"704e278f-04a5-431b-b716-e3a2e71afc10","Type":"ContainerStarted","Data":"6a26e598f27407ae7917d488b9cf63982c313d428fc20c281ced2158430dbbb9"} Dec 06 04:13:25 crc kubenswrapper[4718]: I1206 04:13:25.177768 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb"] Dec 06 04:13:25 crc kubenswrapper[4718]: I1206 04:13:25.180627 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6cdffcbc8f-m28tb"] Dec 06 04:13:25 crc kubenswrapper[4718]: I1206 04:13:25.336995 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bbfde19-028f-469a-8af1-e2b28334b256" path="/var/lib/kubelet/pods/4bbfde19-028f-469a-8af1-e2b28334b256/volumes" Dec 06 04:13:26 crc kubenswrapper[4718]: I1206 04:13:26.154534 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" event={"ID":"704e278f-04a5-431b-b716-e3a2e71afc10","Type":"ContainerStarted","Data":"0eb0ae60f596bf817c4e7cba46ec48b1bdf721529c5bef74c825e98011abc937"} Dec 06 04:13:26 crc kubenswrapper[4718]: I1206 04:13:26.157343 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:26 crc kubenswrapper[4718]: I1206 04:13:26.167008 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" Dec 06 04:13:26 crc kubenswrapper[4718]: I1206 04:13:26.195907 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b7b55f4df-j9qjj" podStartSLOduration=3.195882065 podStartE2EDuration="3.195882065s" podCreationTimestamp="2025-12-06 04:13:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:13:26.179380317 +0000 UTC m=+395.185085508" watchObservedRunningTime="2025-12-06 04:13:26.195882065 +0000 UTC m=+395.201587236" Dec 06 04:13:27 crc kubenswrapper[4718]: I1206 04:13:27.877117 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:13:27 crc kubenswrapper[4718]: I1206 04:13:27.877220 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:13:33 crc kubenswrapper[4718]: I1206 04:13:33.559312 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-fbpfb" Dec 06 04:13:33 crc kubenswrapper[4718]: I1206 04:13:33.631632 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sghf9"] Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.106676 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lnxxq"] Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.108281 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lnxxq" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerName="registry-server" containerID="cri-o://5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0" gracePeriod=30 Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.121078 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwxl5"] Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.121470 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwxl5" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerName="registry-server" containerID="cri-o://1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353" gracePeriod=30 Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.137631 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6kwkv"] Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.138047 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" podUID="3b0daeef-042f-4c26-b228-f879326917b0" containerName="marketplace-operator" containerID="cri-o://522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f" gracePeriod=30 Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.154379 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkmtd"] Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.154715 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nkmtd" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerName="registry-server" containerID="cri-o://569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177" gracePeriod=30 Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.168648 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-47mm6"] Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.168928 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-47mm6" podUID="b5790280-687b-4e83-addd-504672823e57" containerName="registry-server" containerID="cri-o://370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e" gracePeriod=30 Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.173153 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-42x7c"] Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.174012 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.186976 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-42x7c"] Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.203063 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqgsb\" (UniqueName: \"kubernetes.io/projected/49e186cc-90d7-4d3e-973b-1727f13a1c6d-kube-api-access-nqgsb\") pod \"marketplace-operator-79b997595-42x7c\" (UID: \"49e186cc-90d7-4d3e-973b-1727f13a1c6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.203139 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/49e186cc-90d7-4d3e-973b-1727f13a1c6d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-42x7c\" (UID: \"49e186cc-90d7-4d3e-973b-1727f13a1c6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.203524 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49e186cc-90d7-4d3e-973b-1727f13a1c6d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-42x7c\" (UID: \"49e186cc-90d7-4d3e-973b-1727f13a1c6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.307667 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49e186cc-90d7-4d3e-973b-1727f13a1c6d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-42x7c\" (UID: \"49e186cc-90d7-4d3e-973b-1727f13a1c6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.308380 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqgsb\" (UniqueName: \"kubernetes.io/projected/49e186cc-90d7-4d3e-973b-1727f13a1c6d-kube-api-access-nqgsb\") pod \"marketplace-operator-79b997595-42x7c\" (UID: \"49e186cc-90d7-4d3e-973b-1727f13a1c6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.308412 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/49e186cc-90d7-4d3e-973b-1727f13a1c6d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-42x7c\" (UID: \"49e186cc-90d7-4d3e-973b-1727f13a1c6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.309626 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49e186cc-90d7-4d3e-973b-1727f13a1c6d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-42x7c\" (UID: \"49e186cc-90d7-4d3e-973b-1727f13a1c6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.323006 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/49e186cc-90d7-4d3e-973b-1727f13a1c6d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-42x7c\" (UID: \"49e186cc-90d7-4d3e-973b-1727f13a1c6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.333881 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqgsb\" (UniqueName: \"kubernetes.io/projected/49e186cc-90d7-4d3e-973b-1727f13a1c6d-kube-api-access-nqgsb\") pod \"marketplace-operator-79b997595-42x7c\" (UID: \"49e186cc-90d7-4d3e-973b-1727f13a1c6d\") " pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.640981 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.643336 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.726163 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-utilities\") pod \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.726395 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-catalog-content\") pod \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.726494 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7l2t\" (UniqueName: \"kubernetes.io/projected/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-kube-api-access-c7l2t\") pod \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\" (UID: \"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.727397 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-utilities" (OuterVolumeSpecName: "utilities") pod "b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" (UID: "b5ff337f-64c3-437b-a5cd-06dcf7cd96e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.737204 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-kube-api-access-c7l2t" (OuterVolumeSpecName: "kube-api-access-c7l2t") pod "b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" (UID: "b5ff337f-64c3-437b-a5cd-06dcf7cd96e8"). InnerVolumeSpecName "kube-api-access-c7l2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.786159 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" (UID: "b5ff337f-64c3-437b-a5cd-06dcf7cd96e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.818631 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.843848 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.844015 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.844027 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7l2t\" (UniqueName: \"kubernetes.io/projected/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8-kube-api-access-c7l2t\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.868638 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.870967 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.944580 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-utilities\") pod \"b5790280-687b-4e83-addd-504672823e57\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.944661 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-utilities\") pod \"a40acf1b-8c0d-4617-8f41-f48f097cce72\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.944692 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zqwb\" (UniqueName: \"kubernetes.io/projected/b5790280-687b-4e83-addd-504672823e57-kube-api-access-8zqwb\") pod \"b5790280-687b-4e83-addd-504672823e57\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.944765 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwctz\" (UniqueName: \"kubernetes.io/projected/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-kube-api-access-kwctz\") pod \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.944827 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-utilities\") pod \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.944906 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-catalog-content\") pod \"b5790280-687b-4e83-addd-504672823e57\" (UID: \"b5790280-687b-4e83-addd-504672823e57\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.944932 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5x8n\" (UniqueName: \"kubernetes.io/projected/a40acf1b-8c0d-4617-8f41-f48f097cce72-kube-api-access-j5x8n\") pod \"a40acf1b-8c0d-4617-8f41-f48f097cce72\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.944978 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-catalog-content\") pod \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\" (UID: \"989fbd15-c5fb-41b3-bdc8-2c5399b1424a\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.945003 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-catalog-content\") pod \"a40acf1b-8c0d-4617-8f41-f48f097cce72\" (UID: \"a40acf1b-8c0d-4617-8f41-f48f097cce72\") " Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.946372 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-utilities" (OuterVolumeSpecName: "utilities") pod "a40acf1b-8c0d-4617-8f41-f48f097cce72" (UID: "a40acf1b-8c0d-4617-8f41-f48f097cce72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.947911 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-utilities" (OuterVolumeSpecName: "utilities") pod "989fbd15-c5fb-41b3-bdc8-2c5399b1424a" (UID: "989fbd15-c5fb-41b3-bdc8-2c5399b1424a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.950185 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-utilities" (OuterVolumeSpecName: "utilities") pod "b5790280-687b-4e83-addd-504672823e57" (UID: "b5790280-687b-4e83-addd-504672823e57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.957402 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-kube-api-access-kwctz" (OuterVolumeSpecName: "kube-api-access-kwctz") pod "989fbd15-c5fb-41b3-bdc8-2c5399b1424a" (UID: "989fbd15-c5fb-41b3-bdc8-2c5399b1424a"). InnerVolumeSpecName "kube-api-access-kwctz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.957444 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5790280-687b-4e83-addd-504672823e57-kube-api-access-8zqwb" (OuterVolumeSpecName: "kube-api-access-8zqwb") pod "b5790280-687b-4e83-addd-504672823e57" (UID: "b5790280-687b-4e83-addd-504672823e57"). InnerVolumeSpecName "kube-api-access-8zqwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.963222 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a40acf1b-8c0d-4617-8f41-f48f097cce72-kube-api-access-j5x8n" (OuterVolumeSpecName: "kube-api-access-j5x8n") pod "a40acf1b-8c0d-4617-8f41-f48f097cce72" (UID: "a40acf1b-8c0d-4617-8f41-f48f097cce72"). InnerVolumeSpecName "kube-api-access-j5x8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:35 crc kubenswrapper[4718]: I1206 04:13:35.985473 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a40acf1b-8c0d-4617-8f41-f48f097cce72" (UID: "a40acf1b-8c0d-4617-8f41-f48f097cce72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.019206 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "989fbd15-c5fb-41b3-bdc8-2c5399b1424a" (UID: "989fbd15-c5fb-41b3-bdc8-2c5399b1424a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.046520 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwctz\" (UniqueName: \"kubernetes.io/projected/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-kube-api-access-kwctz\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.046568 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.046584 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5x8n\" (UniqueName: \"kubernetes.io/projected/a40acf1b-8c0d-4617-8f41-f48f097cce72-kube-api-access-j5x8n\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.046596 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/989fbd15-c5fb-41b3-bdc8-2c5399b1424a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.046608 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.046619 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.046628 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40acf1b-8c0d-4617-8f41-f48f097cce72-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.046640 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zqwb\" (UniqueName: \"kubernetes.io/projected/b5790280-687b-4e83-addd-504672823e57-kube-api-access-8zqwb\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.062316 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5790280-687b-4e83-addd-504672823e57" (UID: "b5790280-687b-4e83-addd-504672823e57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.116048 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-42x7c"] Dec 06 04:13:36 crc kubenswrapper[4718]: W1206 04:13:36.121418 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49e186cc_90d7_4d3e_973b_1727f13a1c6d.slice/crio-7337ab6f2b69ea70eaf1d1ea68b535cca55784cf15ee05a573fede7e194e4bb6 WatchSource:0}: Error finding container 7337ab6f2b69ea70eaf1d1ea68b535cca55784cf15ee05a573fede7e194e4bb6: Status 404 returned error can't find the container with id 7337ab6f2b69ea70eaf1d1ea68b535cca55784cf15ee05a573fede7e194e4bb6 Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.150317 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5790280-687b-4e83-addd-504672823e57-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.160311 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.214644 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" event={"ID":"49e186cc-90d7-4d3e-973b-1727f13a1c6d","Type":"ContainerStarted","Data":"7337ab6f2b69ea70eaf1d1ea68b535cca55784cf15ee05a573fede7e194e4bb6"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.216694 4718 generic.go:334] "Generic (PLEG): container finished" podID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerID="1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353" exitCode=0 Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.216875 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwxl5" event={"ID":"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8","Type":"ContainerDied","Data":"1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.216922 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwxl5" event={"ID":"b5ff337f-64c3-437b-a5cd-06dcf7cd96e8","Type":"ContainerDied","Data":"59ebc6d50b095b89b78ca8e26d9c395b96dde70a2975a96ca137fae2a3ad4426"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.216944 4718 scope.go:117] "RemoveContainer" containerID="1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.217095 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwxl5" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.220574 4718 generic.go:334] "Generic (PLEG): container finished" podID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerID="569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177" exitCode=0 Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.220633 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkmtd" event={"ID":"a40acf1b-8c0d-4617-8f41-f48f097cce72","Type":"ContainerDied","Data":"569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.220658 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkmtd" event={"ID":"a40acf1b-8c0d-4617-8f41-f48f097cce72","Type":"ContainerDied","Data":"6b1a4ae1e4d4cdc0ac646cfd3817def4f7a997369b19e324d2df1865ece105b0"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.220718 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkmtd" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.222403 4718 generic.go:334] "Generic (PLEG): container finished" podID="3b0daeef-042f-4c26-b228-f879326917b0" containerID="522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f" exitCode=0 Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.222483 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.222689 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" event={"ID":"3b0daeef-042f-4c26-b228-f879326917b0","Type":"ContainerDied","Data":"522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.222710 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6kwkv" event={"ID":"3b0daeef-042f-4c26-b228-f879326917b0","Type":"ContainerDied","Data":"80dcc669aaf712a7083fd6a323cce17cd033077021893156609520e10ae74fe0"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.224963 4718 generic.go:334] "Generic (PLEG): container finished" podID="b5790280-687b-4e83-addd-504672823e57" containerID="370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e" exitCode=0 Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.225041 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-47mm6" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.225130 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47mm6" event={"ID":"b5790280-687b-4e83-addd-504672823e57","Type":"ContainerDied","Data":"370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.225186 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47mm6" event={"ID":"b5790280-687b-4e83-addd-504672823e57","Type":"ContainerDied","Data":"402586121dcfe4de78cc88c9b6afc12b86612d8389a63e2257f3a0b830cc11b3"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.228191 4718 generic.go:334] "Generic (PLEG): container finished" podID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerID="5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0" exitCode=0 Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.228529 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnxxq" event={"ID":"989fbd15-c5fb-41b3-bdc8-2c5399b1424a","Type":"ContainerDied","Data":"5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.228553 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnxxq" event={"ID":"989fbd15-c5fb-41b3-bdc8-2c5399b1424a","Type":"ContainerDied","Data":"7739b998f987ee4bdb3a78b3c6acbe580cb4d62eaf457dc9657f2db827d70c61"} Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.228626 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnxxq" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.237269 4718 scope.go:117] "RemoveContainer" containerID="b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.251166 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-operator-metrics\") pod \"3b0daeef-042f-4c26-b228-f879326917b0\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.251222 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-trusted-ca\") pod \"3b0daeef-042f-4c26-b228-f879326917b0\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.251287 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85lss\" (UniqueName: \"kubernetes.io/projected/3b0daeef-042f-4c26-b228-f879326917b0-kube-api-access-85lss\") pod \"3b0daeef-042f-4c26-b228-f879326917b0\" (UID: \"3b0daeef-042f-4c26-b228-f879326917b0\") " Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.253884 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "3b0daeef-042f-4c26-b228-f879326917b0" (UID: "3b0daeef-042f-4c26-b228-f879326917b0"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.263172 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0daeef-042f-4c26-b228-f879326917b0-kube-api-access-85lss" (OuterVolumeSpecName: "kube-api-access-85lss") pod "3b0daeef-042f-4c26-b228-f879326917b0" (UID: "3b0daeef-042f-4c26-b228-f879326917b0"). InnerVolumeSpecName "kube-api-access-85lss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.271124 4718 scope.go:117] "RemoveContainer" containerID="57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.275396 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "3b0daeef-042f-4c26-b228-f879326917b0" (UID: "3b0daeef-042f-4c26-b228-f879326917b0"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.281860 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwxl5"] Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.290171 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwxl5"] Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.294482 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-47mm6"] Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.298594 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-47mm6"] Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.303818 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkmtd"] Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.308336 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkmtd"] Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.311292 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lnxxq"] Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.313907 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lnxxq"] Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.325338 4718 scope.go:117] "RemoveContainer" containerID="1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.325950 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353\": container with ID starting with 1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353 not found: ID does not exist" containerID="1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.326015 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353"} err="failed to get container status \"1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353\": rpc error: code = NotFound desc = could not find container \"1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353\": container with ID starting with 1a641f0f24ceb4bb6dc95821d9ec35d776172c5bac2e1133dd91e90db6f34353 not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.326051 4718 scope.go:117] "RemoveContainer" containerID="b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.326543 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2\": container with ID starting with b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2 not found: ID does not exist" containerID="b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.326585 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2"} err="failed to get container status \"b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2\": rpc error: code = NotFound desc = could not find container \"b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2\": container with ID starting with b35f7ceff474d7c9d124d047abeccb99b7b71a8d717efae8b92be197829e09c2 not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.326628 4718 scope.go:117] "RemoveContainer" containerID="57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.327094 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200\": container with ID starting with 57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200 not found: ID does not exist" containerID="57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.327159 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200"} err="failed to get container status \"57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200\": rpc error: code = NotFound desc = could not find container \"57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200\": container with ID starting with 57e7e505867ef51b69080713d055d583e71202f49dfa93c7bc421e9188785200 not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.327179 4718 scope.go:117] "RemoveContainer" containerID="569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.339151 4718 scope.go:117] "RemoveContainer" containerID="84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.352150 4718 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.352180 4718 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b0daeef-042f-4c26-b228-f879326917b0-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.352191 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85lss\" (UniqueName: \"kubernetes.io/projected/3b0daeef-042f-4c26-b228-f879326917b0-kube-api-access-85lss\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.352980 4718 scope.go:117] "RemoveContainer" containerID="a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.371952 4718 scope.go:117] "RemoveContainer" containerID="569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.372537 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177\": container with ID starting with 569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177 not found: ID does not exist" containerID="569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.372582 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177"} err="failed to get container status \"569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177\": rpc error: code = NotFound desc = could not find container \"569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177\": container with ID starting with 569d42d3cc8de803108464c2af8db6861e45b58f8e830a9b3e31907eee6d1177 not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.372617 4718 scope.go:117] "RemoveContainer" containerID="84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.372974 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7\": container with ID starting with 84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7 not found: ID does not exist" containerID="84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.373004 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7"} err="failed to get container status \"84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7\": rpc error: code = NotFound desc = could not find container \"84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7\": container with ID starting with 84e3e7ff90af33e32157c1a8346de987121107e3ce0372c20f90d2e651b1b0a7 not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.373023 4718 scope.go:117] "RemoveContainer" containerID="a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.373316 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468\": container with ID starting with a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468 not found: ID does not exist" containerID="a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.373345 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468"} err="failed to get container status \"a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468\": rpc error: code = NotFound desc = could not find container \"a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468\": container with ID starting with a43ff8e7c573cb3b01ec6237a5b90bbde7124c166a9999599b4cf6d9caf95468 not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.373362 4718 scope.go:117] "RemoveContainer" containerID="522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.396184 4718 scope.go:117] "RemoveContainer" containerID="522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.396747 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f\": container with ID starting with 522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f not found: ID does not exist" containerID="522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.396782 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f"} err="failed to get container status \"522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f\": rpc error: code = NotFound desc = could not find container \"522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f\": container with ID starting with 522d0cbaec3ede38b0b0dd5574b8838141da283b65583c7ec8cf3459e4bcac3f not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.396807 4718 scope.go:117] "RemoveContainer" containerID="370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.408680 4718 scope.go:117] "RemoveContainer" containerID="7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.423102 4718 scope.go:117] "RemoveContainer" containerID="691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.446441 4718 scope.go:117] "RemoveContainer" containerID="370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.446996 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e\": container with ID starting with 370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e not found: ID does not exist" containerID="370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.447038 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e"} err="failed to get container status \"370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e\": rpc error: code = NotFound desc = could not find container \"370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e\": container with ID starting with 370af8092f9bb3f735db6cc9186ff3314b45e8d8f4a607049497edc82322e27e not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.447069 4718 scope.go:117] "RemoveContainer" containerID="7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.447568 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2\": container with ID starting with 7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2 not found: ID does not exist" containerID="7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.447616 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2"} err="failed to get container status \"7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2\": rpc error: code = NotFound desc = could not find container \"7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2\": container with ID starting with 7b7c99f894ab600c8f3f87dddbfd91539834a49a693eac5f4a46efbc5a7ac0c2 not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.447645 4718 scope.go:117] "RemoveContainer" containerID="691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.448150 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe\": container with ID starting with 691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe not found: ID does not exist" containerID="691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.448190 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe"} err="failed to get container status \"691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe\": rpc error: code = NotFound desc = could not find container \"691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe\": container with ID starting with 691f2673c186251b679df616e9ebc4059fb79eaa1f7857b280163db2d86ecbfe not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.448216 4718 scope.go:117] "RemoveContainer" containerID="5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.462882 4718 scope.go:117] "RemoveContainer" containerID="9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.474949 4718 scope.go:117] "RemoveContainer" containerID="0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.486522 4718 scope.go:117] "RemoveContainer" containerID="5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.486920 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0\": container with ID starting with 5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0 not found: ID does not exist" containerID="5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.486952 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0"} err="failed to get container status \"5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0\": rpc error: code = NotFound desc = could not find container \"5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0\": container with ID starting with 5484c8d6eb566498676f028604babf60e141d0f184a5883b216c96c0ee3d95f0 not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.486973 4718 scope.go:117] "RemoveContainer" containerID="9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.487267 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049\": container with ID starting with 9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049 not found: ID does not exist" containerID="9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.487293 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049"} err="failed to get container status \"9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049\": rpc error: code = NotFound desc = could not find container \"9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049\": container with ID starting with 9e8e94c2aa6cb4081fd25a7822ef2a2a0e1d9735aa70f869ce673e2ff1d5b049 not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.487306 4718 scope.go:117] "RemoveContainer" containerID="0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce" Dec 06 04:13:36 crc kubenswrapper[4718]: E1206 04:13:36.487679 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce\": container with ID starting with 0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce not found: ID does not exist" containerID="0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.487704 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce"} err="failed to get container status \"0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce\": rpc error: code = NotFound desc = could not find container \"0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce\": container with ID starting with 0db16d9bdbfa198efc232c040048a17dccd20d9dc521b6a35a44f54f728585ce not found: ID does not exist" Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.554136 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6kwkv"] Dec 06 04:13:36 crc kubenswrapper[4718]: I1206 04:13:36.557902 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6kwkv"] Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.236275 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" event={"ID":"49e186cc-90d7-4d3e-973b-1727f13a1c6d","Type":"ContainerStarted","Data":"697168ee6f6bf3b546fee72ec2386487e6ac578e7b5f5f12815d6f4f7a344d6d"} Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.238858 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.241225 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.260976 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-42x7c" podStartSLOduration=2.260958514 podStartE2EDuration="2.260958514s" podCreationTimestamp="2025-12-06 04:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:13:37.255013189 +0000 UTC m=+406.260718350" watchObservedRunningTime="2025-12-06 04:13:37.260958514 +0000 UTC m=+406.266663675" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.319728 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bp8fr"] Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320276 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5790280-687b-4e83-addd-504672823e57" containerName="extract-utilities" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320291 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5790280-687b-4e83-addd-504672823e57" containerName="extract-utilities" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320302 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5790280-687b-4e83-addd-504672823e57" containerName="extract-content" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320309 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5790280-687b-4e83-addd-504672823e57" containerName="extract-content" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320319 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerName="extract-utilities" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320327 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerName="extract-utilities" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320336 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320343 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320353 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerName="extract-content" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320360 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerName="extract-content" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320371 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerName="extract-utilities" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320377 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerName="extract-utilities" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320385 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerName="extract-content" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320391 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerName="extract-content" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320399 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320407 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320419 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320426 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320433 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5790280-687b-4e83-addd-504672823e57" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320439 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5790280-687b-4e83-addd-504672823e57" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320447 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerName="extract-content" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320454 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerName="extract-content" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320468 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0daeef-042f-4c26-b228-f879326917b0" containerName="marketplace-operator" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320474 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0daeef-042f-4c26-b228-f879326917b0" containerName="marketplace-operator" Dec 06 04:13:37 crc kubenswrapper[4718]: E1206 04:13:37.320486 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerName="extract-utilities" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320493 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerName="extract-utilities" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320584 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b0daeef-042f-4c26-b228-f879326917b0" containerName="marketplace-operator" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320599 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320607 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5790280-687b-4e83-addd-504672823e57" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320618 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.320625 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" containerName="registry-server" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.321483 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.323453 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.334203 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b0daeef-042f-4c26-b228-f879326917b0" path="/var/lib/kubelet/pods/3b0daeef-042f-4c26-b228-f879326917b0/volumes" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.334989 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="989fbd15-c5fb-41b3-bdc8-2c5399b1424a" path="/var/lib/kubelet/pods/989fbd15-c5fb-41b3-bdc8-2c5399b1424a/volumes" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.342392 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a40acf1b-8c0d-4617-8f41-f48f097cce72" path="/var/lib/kubelet/pods/a40acf1b-8c0d-4617-8f41-f48f097cce72/volumes" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.343375 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5790280-687b-4e83-addd-504672823e57" path="/var/lib/kubelet/pods/b5790280-687b-4e83-addd-504672823e57/volumes" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.344181 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5ff337f-64c3-437b-a5cd-06dcf7cd96e8" path="/var/lib/kubelet/pods/b5ff337f-64c3-437b-a5cd-06dcf7cd96e8/volumes" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.345420 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bp8fr"] Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.466076 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15cc3c85-bfb7-4f19-9498-840989a76366-catalog-content\") pod \"redhat-marketplace-bp8fr\" (UID: \"15cc3c85-bfb7-4f19-9498-840989a76366\") " pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.466121 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl4f7\" (UniqueName: \"kubernetes.io/projected/15cc3c85-bfb7-4f19-9498-840989a76366-kube-api-access-gl4f7\") pod \"redhat-marketplace-bp8fr\" (UID: \"15cc3c85-bfb7-4f19-9498-840989a76366\") " pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.466178 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15cc3c85-bfb7-4f19-9498-840989a76366-utilities\") pod \"redhat-marketplace-bp8fr\" (UID: \"15cc3c85-bfb7-4f19-9498-840989a76366\") " pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.524390 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j784j"] Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.525513 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.528395 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.535359 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j784j"] Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.567849 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15cc3c85-bfb7-4f19-9498-840989a76366-catalog-content\") pod \"redhat-marketplace-bp8fr\" (UID: \"15cc3c85-bfb7-4f19-9498-840989a76366\") " pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.567898 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl4f7\" (UniqueName: \"kubernetes.io/projected/15cc3c85-bfb7-4f19-9498-840989a76366-kube-api-access-gl4f7\") pod \"redhat-marketplace-bp8fr\" (UID: \"15cc3c85-bfb7-4f19-9498-840989a76366\") " pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.567951 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15cc3c85-bfb7-4f19-9498-840989a76366-utilities\") pod \"redhat-marketplace-bp8fr\" (UID: \"15cc3c85-bfb7-4f19-9498-840989a76366\") " pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.568448 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15cc3c85-bfb7-4f19-9498-840989a76366-catalog-content\") pod \"redhat-marketplace-bp8fr\" (UID: \"15cc3c85-bfb7-4f19-9498-840989a76366\") " pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.568502 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15cc3c85-bfb7-4f19-9498-840989a76366-utilities\") pod \"redhat-marketplace-bp8fr\" (UID: \"15cc3c85-bfb7-4f19-9498-840989a76366\") " pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.592319 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl4f7\" (UniqueName: \"kubernetes.io/projected/15cc3c85-bfb7-4f19-9498-840989a76366-kube-api-access-gl4f7\") pod \"redhat-marketplace-bp8fr\" (UID: \"15cc3c85-bfb7-4f19-9498-840989a76366\") " pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.639307 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.669838 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq5dl\" (UniqueName: \"kubernetes.io/projected/04067a51-e169-4754-b847-e26e29ec8213-kube-api-access-nq5dl\") pod \"redhat-operators-j784j\" (UID: \"04067a51-e169-4754-b847-e26e29ec8213\") " pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.669976 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04067a51-e169-4754-b847-e26e29ec8213-utilities\") pod \"redhat-operators-j784j\" (UID: \"04067a51-e169-4754-b847-e26e29ec8213\") " pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.670032 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04067a51-e169-4754-b847-e26e29ec8213-catalog-content\") pod \"redhat-operators-j784j\" (UID: \"04067a51-e169-4754-b847-e26e29ec8213\") " pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.773839 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04067a51-e169-4754-b847-e26e29ec8213-catalog-content\") pod \"redhat-operators-j784j\" (UID: \"04067a51-e169-4754-b847-e26e29ec8213\") " pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.773898 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq5dl\" (UniqueName: \"kubernetes.io/projected/04067a51-e169-4754-b847-e26e29ec8213-kube-api-access-nq5dl\") pod \"redhat-operators-j784j\" (UID: \"04067a51-e169-4754-b847-e26e29ec8213\") " pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.773948 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04067a51-e169-4754-b847-e26e29ec8213-utilities\") pod \"redhat-operators-j784j\" (UID: \"04067a51-e169-4754-b847-e26e29ec8213\") " pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.774368 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04067a51-e169-4754-b847-e26e29ec8213-utilities\") pod \"redhat-operators-j784j\" (UID: \"04067a51-e169-4754-b847-e26e29ec8213\") " pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.774600 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04067a51-e169-4754-b847-e26e29ec8213-catalog-content\") pod \"redhat-operators-j784j\" (UID: \"04067a51-e169-4754-b847-e26e29ec8213\") " pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.792454 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq5dl\" (UniqueName: \"kubernetes.io/projected/04067a51-e169-4754-b847-e26e29ec8213-kube-api-access-nq5dl\") pod \"redhat-operators-j784j\" (UID: \"04067a51-e169-4754-b847-e26e29ec8213\") " pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:37 crc kubenswrapper[4718]: I1206 04:13:37.879581 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:38 crc kubenswrapper[4718]: I1206 04:13:38.047989 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bp8fr"] Dec 06 04:13:38 crc kubenswrapper[4718]: W1206 04:13:38.051970 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15cc3c85_bfb7_4f19_9498_840989a76366.slice/crio-856af1b044b6d849f18300090c650bb8110749e4a5f03dfa877665c7e6fa3402 WatchSource:0}: Error finding container 856af1b044b6d849f18300090c650bb8110749e4a5f03dfa877665c7e6fa3402: Status 404 returned error can't find the container with id 856af1b044b6d849f18300090c650bb8110749e4a5f03dfa877665c7e6fa3402 Dec 06 04:13:38 crc kubenswrapper[4718]: I1206 04:13:38.247966 4718 generic.go:334] "Generic (PLEG): container finished" podID="15cc3c85-bfb7-4f19-9498-840989a76366" containerID="03f3f1c789eeb2b4e097fd44364e8eeb7489f2470eb753622f4a0c1e87a2a549" exitCode=0 Dec 06 04:13:38 crc kubenswrapper[4718]: I1206 04:13:38.248933 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp8fr" event={"ID":"15cc3c85-bfb7-4f19-9498-840989a76366","Type":"ContainerDied","Data":"03f3f1c789eeb2b4e097fd44364e8eeb7489f2470eb753622f4a0c1e87a2a549"} Dec 06 04:13:38 crc kubenswrapper[4718]: I1206 04:13:38.248958 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp8fr" event={"ID":"15cc3c85-bfb7-4f19-9498-840989a76366","Type":"ContainerStarted","Data":"856af1b044b6d849f18300090c650bb8110749e4a5f03dfa877665c7e6fa3402"} Dec 06 04:13:38 crc kubenswrapper[4718]: I1206 04:13:38.277914 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j784j"] Dec 06 04:13:38 crc kubenswrapper[4718]: W1206 04:13:38.280849 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04067a51_e169_4754_b847_e26e29ec8213.slice/crio-83e398fce473d050f20256419831925af2b583d01f1da548c6f0d4f07f8b79b1 WatchSource:0}: Error finding container 83e398fce473d050f20256419831925af2b583d01f1da548c6f0d4f07f8b79b1: Status 404 returned error can't find the container with id 83e398fce473d050f20256419831925af2b583d01f1da548c6f0d4f07f8b79b1 Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.258729 4718 generic.go:334] "Generic (PLEG): container finished" podID="04067a51-e169-4754-b847-e26e29ec8213" containerID="b2979e2ccbb6f6bdb1a80b5ce19e491ef8e2094f9f08f24eb2259cdff441824d" exitCode=0 Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.258916 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j784j" event={"ID":"04067a51-e169-4754-b847-e26e29ec8213","Type":"ContainerDied","Data":"b2979e2ccbb6f6bdb1a80b5ce19e491ef8e2094f9f08f24eb2259cdff441824d"} Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.259322 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j784j" event={"ID":"04067a51-e169-4754-b847-e26e29ec8213","Type":"ContainerStarted","Data":"83e398fce473d050f20256419831925af2b583d01f1da548c6f0d4f07f8b79b1"} Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.269260 4718 generic.go:334] "Generic (PLEG): container finished" podID="15cc3c85-bfb7-4f19-9498-840989a76366" containerID="b97312a171fa05d5622eb221ef6541e0acb67a3e79a4cc65036727346e3176e1" exitCode=0 Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.269402 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp8fr" event={"ID":"15cc3c85-bfb7-4f19-9498-840989a76366","Type":"ContainerDied","Data":"b97312a171fa05d5622eb221ef6541e0acb67a3e79a4cc65036727346e3176e1"} Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.729853 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2sbp7"] Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.732427 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.735213 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.742956 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2sbp7"] Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.801282 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb623567-41f3-4c02-837a-40255117f3e8-catalog-content\") pod \"community-operators-2sbp7\" (UID: \"fb623567-41f3-4c02-837a-40255117f3e8\") " pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.801334 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb623567-41f3-4c02-837a-40255117f3e8-utilities\") pod \"community-operators-2sbp7\" (UID: \"fb623567-41f3-4c02-837a-40255117f3e8\") " pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.801363 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5frzt\" (UniqueName: \"kubernetes.io/projected/fb623567-41f3-4c02-837a-40255117f3e8-kube-api-access-5frzt\") pod \"community-operators-2sbp7\" (UID: \"fb623567-41f3-4c02-837a-40255117f3e8\") " pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.902746 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb623567-41f3-4c02-837a-40255117f3e8-catalog-content\") pod \"community-operators-2sbp7\" (UID: \"fb623567-41f3-4c02-837a-40255117f3e8\") " pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.902818 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb623567-41f3-4c02-837a-40255117f3e8-utilities\") pod \"community-operators-2sbp7\" (UID: \"fb623567-41f3-4c02-837a-40255117f3e8\") " pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.902854 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5frzt\" (UniqueName: \"kubernetes.io/projected/fb623567-41f3-4c02-837a-40255117f3e8-kube-api-access-5frzt\") pod \"community-operators-2sbp7\" (UID: \"fb623567-41f3-4c02-837a-40255117f3e8\") " pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.903597 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb623567-41f3-4c02-837a-40255117f3e8-catalog-content\") pod \"community-operators-2sbp7\" (UID: \"fb623567-41f3-4c02-837a-40255117f3e8\") " pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.903832 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb623567-41f3-4c02-837a-40255117f3e8-utilities\") pod \"community-operators-2sbp7\" (UID: \"fb623567-41f3-4c02-837a-40255117f3e8\") " pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.926280 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5frzt\" (UniqueName: \"kubernetes.io/projected/fb623567-41f3-4c02-837a-40255117f3e8-kube-api-access-5frzt\") pod \"community-operators-2sbp7\" (UID: \"fb623567-41f3-4c02-837a-40255117f3e8\") " pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.930950 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-clxct"] Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.932364 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.934980 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 04:13:39 crc kubenswrapper[4718]: I1206 04:13:39.935176 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-clxct"] Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.004345 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c27751f-4adc-45e6-8e0f-1a4ead35f894-catalog-content\") pod \"certified-operators-clxct\" (UID: \"1c27751f-4adc-45e6-8e0f-1a4ead35f894\") " pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.004682 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c27751f-4adc-45e6-8e0f-1a4ead35f894-utilities\") pod \"certified-operators-clxct\" (UID: \"1c27751f-4adc-45e6-8e0f-1a4ead35f894\") " pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.004718 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87k9p\" (UniqueName: \"kubernetes.io/projected/1c27751f-4adc-45e6-8e0f-1a4ead35f894-kube-api-access-87k9p\") pod \"certified-operators-clxct\" (UID: \"1c27751f-4adc-45e6-8e0f-1a4ead35f894\") " pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.060498 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.105946 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c27751f-4adc-45e6-8e0f-1a4ead35f894-utilities\") pod \"certified-operators-clxct\" (UID: \"1c27751f-4adc-45e6-8e0f-1a4ead35f894\") " pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.106009 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87k9p\" (UniqueName: \"kubernetes.io/projected/1c27751f-4adc-45e6-8e0f-1a4ead35f894-kube-api-access-87k9p\") pod \"certified-operators-clxct\" (UID: \"1c27751f-4adc-45e6-8e0f-1a4ead35f894\") " pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.106113 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c27751f-4adc-45e6-8e0f-1a4ead35f894-catalog-content\") pod \"certified-operators-clxct\" (UID: \"1c27751f-4adc-45e6-8e0f-1a4ead35f894\") " pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.106656 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c27751f-4adc-45e6-8e0f-1a4ead35f894-utilities\") pod \"certified-operators-clxct\" (UID: \"1c27751f-4adc-45e6-8e0f-1a4ead35f894\") " pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.106670 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c27751f-4adc-45e6-8e0f-1a4ead35f894-catalog-content\") pod \"certified-operators-clxct\" (UID: \"1c27751f-4adc-45e6-8e0f-1a4ead35f894\") " pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.121912 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87k9p\" (UniqueName: \"kubernetes.io/projected/1c27751f-4adc-45e6-8e0f-1a4ead35f894-kube-api-access-87k9p\") pod \"certified-operators-clxct\" (UID: \"1c27751f-4adc-45e6-8e0f-1a4ead35f894\") " pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.268379 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.276746 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j784j" event={"ID":"04067a51-e169-4754-b847-e26e29ec8213","Type":"ContainerStarted","Data":"a50b03f603ee2c0de0f3b84d25ea93bfd5f91aae1fabb0212b0c0bc9746fbedf"} Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.282591 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp8fr" event={"ID":"15cc3c85-bfb7-4f19-9498-840989a76366","Type":"ContainerStarted","Data":"a0226b2a7cb7203125851d81fdeaadb6ce0367a08b2d826ce3b916c415fcdf30"} Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.311654 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bp8fr" podStartSLOduration=1.8372358229999999 podStartE2EDuration="3.311634766s" podCreationTimestamp="2025-12-06 04:13:37 +0000 UTC" firstStartedPulling="2025-12-06 04:13:38.250299115 +0000 UTC m=+407.256004276" lastFinishedPulling="2025-12-06 04:13:39.724698038 +0000 UTC m=+408.730403219" observedRunningTime="2025-12-06 04:13:40.308191121 +0000 UTC m=+409.313896282" watchObservedRunningTime="2025-12-06 04:13:40.311634766 +0000 UTC m=+409.317339927" Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.498972 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2sbp7"] Dec 06 04:13:40 crc kubenswrapper[4718]: W1206 04:13:40.503428 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb623567_41f3_4c02_837a_40255117f3e8.slice/crio-a68551cd03edd0623ffdeea0dbb66b7590426448ad2a307684f0d3d3ca022ba7 WatchSource:0}: Error finding container a68551cd03edd0623ffdeea0dbb66b7590426448ad2a307684f0d3d3ca022ba7: Status 404 returned error can't find the container with id a68551cd03edd0623ffdeea0dbb66b7590426448ad2a307684f0d3d3ca022ba7 Dec 06 04:13:40 crc kubenswrapper[4718]: I1206 04:13:40.665129 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-clxct"] Dec 06 04:13:40 crc kubenswrapper[4718]: W1206 04:13:40.678011 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c27751f_4adc_45e6_8e0f_1a4ead35f894.slice/crio-5a5d34bf3a799f0c1cea79205bf26dc5c1ec102b4f5f3302635b46cc442a4bbc WatchSource:0}: Error finding container 5a5d34bf3a799f0c1cea79205bf26dc5c1ec102b4f5f3302635b46cc442a4bbc: Status 404 returned error can't find the container with id 5a5d34bf3a799f0c1cea79205bf26dc5c1ec102b4f5f3302635b46cc442a4bbc Dec 06 04:13:41 crc kubenswrapper[4718]: I1206 04:13:41.290150 4718 generic.go:334] "Generic (PLEG): container finished" podID="1c27751f-4adc-45e6-8e0f-1a4ead35f894" containerID="f74e6f322b8cacd70881dbd645253ff9da27e07619952e191f6c6d1df435349f" exitCode=0 Dec 06 04:13:41 crc kubenswrapper[4718]: I1206 04:13:41.290220 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clxct" event={"ID":"1c27751f-4adc-45e6-8e0f-1a4ead35f894","Type":"ContainerDied","Data":"f74e6f322b8cacd70881dbd645253ff9da27e07619952e191f6c6d1df435349f"} Dec 06 04:13:41 crc kubenswrapper[4718]: I1206 04:13:41.290268 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clxct" event={"ID":"1c27751f-4adc-45e6-8e0f-1a4ead35f894","Type":"ContainerStarted","Data":"5a5d34bf3a799f0c1cea79205bf26dc5c1ec102b4f5f3302635b46cc442a4bbc"} Dec 06 04:13:41 crc kubenswrapper[4718]: I1206 04:13:41.294550 4718 generic.go:334] "Generic (PLEG): container finished" podID="04067a51-e169-4754-b847-e26e29ec8213" containerID="a50b03f603ee2c0de0f3b84d25ea93bfd5f91aae1fabb0212b0c0bc9746fbedf" exitCode=0 Dec 06 04:13:41 crc kubenswrapper[4718]: I1206 04:13:41.294777 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j784j" event={"ID":"04067a51-e169-4754-b847-e26e29ec8213","Type":"ContainerDied","Data":"a50b03f603ee2c0de0f3b84d25ea93bfd5f91aae1fabb0212b0c0bc9746fbedf"} Dec 06 04:13:41 crc kubenswrapper[4718]: I1206 04:13:41.297657 4718 generic.go:334] "Generic (PLEG): container finished" podID="fb623567-41f3-4c02-837a-40255117f3e8" containerID="a4af52961334875a229bf70baf1ee422b4ac5c0ada3ca38a1afb1b0633d5c6e7" exitCode=0 Dec 06 04:13:41 crc kubenswrapper[4718]: I1206 04:13:41.297742 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sbp7" event={"ID":"fb623567-41f3-4c02-837a-40255117f3e8","Type":"ContainerDied","Data":"a4af52961334875a229bf70baf1ee422b4ac5c0ada3ca38a1afb1b0633d5c6e7"} Dec 06 04:13:41 crc kubenswrapper[4718]: I1206 04:13:41.297819 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sbp7" event={"ID":"fb623567-41f3-4c02-837a-40255117f3e8","Type":"ContainerStarted","Data":"a68551cd03edd0623ffdeea0dbb66b7590426448ad2a307684f0d3d3ca022ba7"} Dec 06 04:13:42 crc kubenswrapper[4718]: I1206 04:13:42.303438 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clxct" event={"ID":"1c27751f-4adc-45e6-8e0f-1a4ead35f894","Type":"ContainerStarted","Data":"bb49c83bae5220d717773d95c296f9c6d86ff1e907ef20ed4f35dff5ebea697e"} Dec 06 04:13:42 crc kubenswrapper[4718]: I1206 04:13:42.306385 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j784j" event={"ID":"04067a51-e169-4754-b847-e26e29ec8213","Type":"ContainerStarted","Data":"b42f0ef507ab25758a661e20bae88620912be793416bc6f9581140e1948890ba"} Dec 06 04:13:42 crc kubenswrapper[4718]: I1206 04:13:42.308405 4718 generic.go:334] "Generic (PLEG): container finished" podID="fb623567-41f3-4c02-837a-40255117f3e8" containerID="3c5ba08fd4afc8fd168dbd2e75f53f747ed363bb11d9fb9447839539a5b14c46" exitCode=0 Dec 06 04:13:42 crc kubenswrapper[4718]: I1206 04:13:42.308437 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sbp7" event={"ID":"fb623567-41f3-4c02-837a-40255117f3e8","Type":"ContainerDied","Data":"3c5ba08fd4afc8fd168dbd2e75f53f747ed363bb11d9fb9447839539a5b14c46"} Dec 06 04:13:42 crc kubenswrapper[4718]: I1206 04:13:42.347166 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j784j" podStartSLOduration=2.921367682 podStartE2EDuration="5.34715094s" podCreationTimestamp="2025-12-06 04:13:37 +0000 UTC" firstStartedPulling="2025-12-06 04:13:39.264282021 +0000 UTC m=+408.269987182" lastFinishedPulling="2025-12-06 04:13:41.690065279 +0000 UTC m=+410.695770440" observedRunningTime="2025-12-06 04:13:42.344754503 +0000 UTC m=+411.350459664" watchObservedRunningTime="2025-12-06 04:13:42.34715094 +0000 UTC m=+411.352856091" Dec 06 04:13:43 crc kubenswrapper[4718]: I1206 04:13:43.315535 4718 generic.go:334] "Generic (PLEG): container finished" podID="1c27751f-4adc-45e6-8e0f-1a4ead35f894" containerID="bb49c83bae5220d717773d95c296f9c6d86ff1e907ef20ed4f35dff5ebea697e" exitCode=0 Dec 06 04:13:43 crc kubenswrapper[4718]: I1206 04:13:43.315601 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clxct" event={"ID":"1c27751f-4adc-45e6-8e0f-1a4ead35f894","Type":"ContainerDied","Data":"bb49c83bae5220d717773d95c296f9c6d86ff1e907ef20ed4f35dff5ebea697e"} Dec 06 04:13:44 crc kubenswrapper[4718]: I1206 04:13:44.322867 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sbp7" event={"ID":"fb623567-41f3-4c02-837a-40255117f3e8","Type":"ContainerStarted","Data":"f1717e167710d8d595c93069dd9aef7d8773c1d33a7cf229a14b5798e656ba6d"} Dec 06 04:13:44 crc kubenswrapper[4718]: I1206 04:13:44.324707 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clxct" event={"ID":"1c27751f-4adc-45e6-8e0f-1a4ead35f894","Type":"ContainerStarted","Data":"bd8b9670b74aac8d7046cd09d490437562233b97e311c64de61e02ddfbb5c74b"} Dec 06 04:13:44 crc kubenswrapper[4718]: I1206 04:13:44.337987 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2sbp7" podStartSLOduration=3.905936065 podStartE2EDuration="5.337972529s" podCreationTimestamp="2025-12-06 04:13:39 +0000 UTC" firstStartedPulling="2025-12-06 04:13:41.299285349 +0000 UTC m=+410.304990510" lastFinishedPulling="2025-12-06 04:13:42.731321813 +0000 UTC m=+411.737026974" observedRunningTime="2025-12-06 04:13:44.337139245 +0000 UTC m=+413.342844406" watchObservedRunningTime="2025-12-06 04:13:44.337972529 +0000 UTC m=+413.343677690" Dec 06 04:13:47 crc kubenswrapper[4718]: I1206 04:13:47.639756 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:47 crc kubenswrapper[4718]: I1206 04:13:47.641169 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:47 crc kubenswrapper[4718]: I1206 04:13:47.689094 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:47 crc kubenswrapper[4718]: I1206 04:13:47.707861 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-clxct" podStartSLOduration=6.187770973 podStartE2EDuration="8.707839915s" podCreationTimestamp="2025-12-06 04:13:39 +0000 UTC" firstStartedPulling="2025-12-06 04:13:41.292302055 +0000 UTC m=+410.298007216" lastFinishedPulling="2025-12-06 04:13:43.812370997 +0000 UTC m=+412.818076158" observedRunningTime="2025-12-06 04:13:44.352553384 +0000 UTC m=+413.358258535" watchObservedRunningTime="2025-12-06 04:13:47.707839915 +0000 UTC m=+416.713545086" Dec 06 04:13:47 crc kubenswrapper[4718]: I1206 04:13:47.880689 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:47 crc kubenswrapper[4718]: I1206 04:13:47.881447 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:47 crc kubenswrapper[4718]: I1206 04:13:47.920147 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:48 crc kubenswrapper[4718]: I1206 04:13:48.380323 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bp8fr" Dec 06 04:13:48 crc kubenswrapper[4718]: I1206 04:13:48.385305 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j784j" Dec 06 04:13:50 crc kubenswrapper[4718]: I1206 04:13:50.061093 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:50 crc kubenswrapper[4718]: I1206 04:13:50.061394 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:50 crc kubenswrapper[4718]: I1206 04:13:50.097349 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:50 crc kubenswrapper[4718]: I1206 04:13:50.268825 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:50 crc kubenswrapper[4718]: I1206 04:13:50.268904 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:50 crc kubenswrapper[4718]: I1206 04:13:50.318813 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:50 crc kubenswrapper[4718]: I1206 04:13:50.393396 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2sbp7" Dec 06 04:13:50 crc kubenswrapper[4718]: I1206 04:13:50.410119 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-clxct" Dec 06 04:13:57 crc kubenswrapper[4718]: I1206 04:13:57.877581 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:13:57 crc kubenswrapper[4718]: I1206 04:13:57.878174 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:13:57 crc kubenswrapper[4718]: I1206 04:13:57.878259 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:13:57 crc kubenswrapper[4718]: I1206 04:13:57.878869 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c4e3740d9b890ba2b04c60617f993f3387fea120a80ab270801f84ab026d7dd7"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:13:57 crc kubenswrapper[4718]: I1206 04:13:57.878943 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://c4e3740d9b890ba2b04c60617f993f3387fea120a80ab270801f84ab026d7dd7" gracePeriod=600 Dec 06 04:13:58 crc kubenswrapper[4718]: I1206 04:13:58.393447 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="c4e3740d9b890ba2b04c60617f993f3387fea120a80ab270801f84ab026d7dd7" exitCode=0 Dec 06 04:13:58 crc kubenswrapper[4718]: I1206 04:13:58.393547 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"c4e3740d9b890ba2b04c60617f993f3387fea120a80ab270801f84ab026d7dd7"} Dec 06 04:13:58 crc kubenswrapper[4718]: I1206 04:13:58.393770 4718 scope.go:117] "RemoveContainer" containerID="1e89cfad69d134d0ea314a5b18b9cbbd5d5b99e966e56aee9c52c2893d41c109" Dec 06 04:13:58 crc kubenswrapper[4718]: I1206 04:13:58.696092 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" podUID="b47eec75-b43c-486d-be96-dce4faceddbf" containerName="registry" containerID="cri-o://39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36" gracePeriod=30 Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.169426 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.265086 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqxc2\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-kube-api-access-fqxc2\") pod \"b47eec75-b43c-486d-be96-dce4faceddbf\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.265153 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-trusted-ca\") pod \"b47eec75-b43c-486d-be96-dce4faceddbf\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.265200 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-bound-sa-token\") pod \"b47eec75-b43c-486d-be96-dce4faceddbf\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.265262 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b47eec75-b43c-486d-be96-dce4faceddbf-ca-trust-extracted\") pod \"b47eec75-b43c-486d-be96-dce4faceddbf\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.265430 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"b47eec75-b43c-486d-be96-dce4faceddbf\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.265458 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-registry-certificates\") pod \"b47eec75-b43c-486d-be96-dce4faceddbf\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.265476 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-registry-tls\") pod \"b47eec75-b43c-486d-be96-dce4faceddbf\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.265518 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b47eec75-b43c-486d-be96-dce4faceddbf-installation-pull-secrets\") pod \"b47eec75-b43c-486d-be96-dce4faceddbf\" (UID: \"b47eec75-b43c-486d-be96-dce4faceddbf\") " Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.266580 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "b47eec75-b43c-486d-be96-dce4faceddbf" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.266656 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "b47eec75-b43c-486d-be96-dce4faceddbf" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.273246 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-kube-api-access-fqxc2" (OuterVolumeSpecName: "kube-api-access-fqxc2") pod "b47eec75-b43c-486d-be96-dce4faceddbf" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf"). InnerVolumeSpecName "kube-api-access-fqxc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.273281 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "b47eec75-b43c-486d-be96-dce4faceddbf" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.274184 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "b47eec75-b43c-486d-be96-dce4faceddbf" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.274647 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "b47eec75-b43c-486d-be96-dce4faceddbf" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.275774 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b47eec75-b43c-486d-be96-dce4faceddbf-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "b47eec75-b43c-486d-be96-dce4faceddbf" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.287064 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b47eec75-b43c-486d-be96-dce4faceddbf-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "b47eec75-b43c-486d-be96-dce4faceddbf" (UID: "b47eec75-b43c-486d-be96-dce4faceddbf"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.367197 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqxc2\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-kube-api-access-fqxc2\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.367453 4718 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.367548 4718 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.367609 4718 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b47eec75-b43c-486d-be96-dce4faceddbf-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.367724 4718 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b47eec75-b43c-486d-be96-dce4faceddbf-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.367787 4718 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b47eec75-b43c-486d-be96-dce4faceddbf-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.367842 4718 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b47eec75-b43c-486d-be96-dce4faceddbf-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.400177 4718 generic.go:334] "Generic (PLEG): container finished" podID="b47eec75-b43c-486d-be96-dce4faceddbf" containerID="39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36" exitCode=0 Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.400250 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" event={"ID":"b47eec75-b43c-486d-be96-dce4faceddbf","Type":"ContainerDied","Data":"39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36"} Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.400305 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" event={"ID":"b47eec75-b43c-486d-be96-dce4faceddbf","Type":"ContainerDied","Data":"30d4fd3eefd2835a0e0684928307251fd68f1c35b0a22f12d87830925962f7b4"} Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.400342 4718 scope.go:117] "RemoveContainer" containerID="39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.400850 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sghf9" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.403608 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"ab28e895cdd83b428eae8d7bd3a859b7eb41e3f8218509581d34db22f8db6925"} Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.420221 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sghf9"] Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.424842 4718 scope.go:117] "RemoveContainer" containerID="39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36" Dec 06 04:13:59 crc kubenswrapper[4718]: E1206 04:13:59.425774 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36\": container with ID starting with 39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36 not found: ID does not exist" containerID="39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.425907 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36"} err="failed to get container status \"39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36\": rpc error: code = NotFound desc = could not find container \"39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36\": container with ID starting with 39af9465a0599567936c0f53080e4a16240cd8cf30f223bdf901d272b3e8fc36 not found: ID does not exist" Dec 06 04:13:59 crc kubenswrapper[4718]: I1206 04:13:59.429246 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sghf9"] Dec 06 04:14:01 crc kubenswrapper[4718]: I1206 04:14:01.340565 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b47eec75-b43c-486d-be96-dce4faceddbf" path="/var/lib/kubelet/pods/b47eec75-b43c-486d-be96-dce4faceddbf/volumes" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.168669 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx"] Dec 06 04:15:00 crc kubenswrapper[4718]: E1206 04:15:00.169500 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b47eec75-b43c-486d-be96-dce4faceddbf" containerName="registry" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.169516 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b47eec75-b43c-486d-be96-dce4faceddbf" containerName="registry" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.169675 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="b47eec75-b43c-486d-be96-dce4faceddbf" containerName="registry" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.170115 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.172665 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.172707 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.181985 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx"] Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.266091 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-secret-volume\") pod \"collect-profiles-29416575-njwhx\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.266180 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-config-volume\") pod \"collect-profiles-29416575-njwhx\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.266279 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5q9r\" (UniqueName: \"kubernetes.io/projected/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-kube-api-access-k5q9r\") pod \"collect-profiles-29416575-njwhx\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.367412 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-config-volume\") pod \"collect-profiles-29416575-njwhx\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.367490 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5q9r\" (UniqueName: \"kubernetes.io/projected/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-kube-api-access-k5q9r\") pod \"collect-profiles-29416575-njwhx\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.367533 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-secret-volume\") pod \"collect-profiles-29416575-njwhx\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.368736 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-config-volume\") pod \"collect-profiles-29416575-njwhx\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.377169 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-secret-volume\") pod \"collect-profiles-29416575-njwhx\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.393908 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5q9r\" (UniqueName: \"kubernetes.io/projected/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-kube-api-access-k5q9r\") pod \"collect-profiles-29416575-njwhx\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.492541 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:00 crc kubenswrapper[4718]: I1206 04:15:00.974376 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx"] Dec 06 04:15:00 crc kubenswrapper[4718]: W1206 04:15:00.979442 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb8d8d9_1e48_4bac_a3f8_e5da23b983c5.slice/crio-656c2bd50eef302666daee4f337ee604e96425471180bf5bed40fc566af88a25 WatchSource:0}: Error finding container 656c2bd50eef302666daee4f337ee604e96425471180bf5bed40fc566af88a25: Status 404 returned error can't find the container with id 656c2bd50eef302666daee4f337ee604e96425471180bf5bed40fc566af88a25 Dec 06 04:15:01 crc kubenswrapper[4718]: I1206 04:15:01.804744 4718 generic.go:334] "Generic (PLEG): container finished" podID="cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5" containerID="3b7dc7d0a513b433ef261fe6134854d234bb201d2e33846d3acf1c012fc408b0" exitCode=0 Dec 06 04:15:01 crc kubenswrapper[4718]: I1206 04:15:01.804917 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" event={"ID":"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5","Type":"ContainerDied","Data":"3b7dc7d0a513b433ef261fe6134854d234bb201d2e33846d3acf1c012fc408b0"} Dec 06 04:15:01 crc kubenswrapper[4718]: I1206 04:15:01.805158 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" event={"ID":"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5","Type":"ContainerStarted","Data":"656c2bd50eef302666daee4f337ee604e96425471180bf5bed40fc566af88a25"} Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.092968 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.206020 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5q9r\" (UniqueName: \"kubernetes.io/projected/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-kube-api-access-k5q9r\") pod \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.206169 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-secret-volume\") pod \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.206295 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-config-volume\") pod \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\" (UID: \"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5\") " Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.207758 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-config-volume" (OuterVolumeSpecName: "config-volume") pod "cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5" (UID: "cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.212039 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-kube-api-access-k5q9r" (OuterVolumeSpecName: "kube-api-access-k5q9r") pod "cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5" (UID: "cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5"). InnerVolumeSpecName "kube-api-access-k5q9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.212166 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5" (UID: "cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.308082 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5q9r\" (UniqueName: \"kubernetes.io/projected/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-kube-api-access-k5q9r\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.308131 4718 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.308147 4718 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.818614 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" event={"ID":"cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5","Type":"ContainerDied","Data":"656c2bd50eef302666daee4f337ee604e96425471180bf5bed40fc566af88a25"} Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.818665 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="656c2bd50eef302666daee4f337ee604e96425471180bf5bed40fc566af88a25" Dec 06 04:15:03 crc kubenswrapper[4718]: I1206 04:15:03.818671 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416575-njwhx" Dec 06 04:16:27 crc kubenswrapper[4718]: I1206 04:16:27.876657 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:16:27 crc kubenswrapper[4718]: I1206 04:16:27.877178 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:16:57 crc kubenswrapper[4718]: I1206 04:16:57.877672 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:16:57 crc kubenswrapper[4718]: I1206 04:16:57.878123 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:17:27 crc kubenswrapper[4718]: I1206 04:17:27.877478 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:17:27 crc kubenswrapper[4718]: I1206 04:17:27.878159 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:17:27 crc kubenswrapper[4718]: I1206 04:17:27.878224 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:17:27 crc kubenswrapper[4718]: I1206 04:17:27.879042 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab28e895cdd83b428eae8d7bd3a859b7eb41e3f8218509581d34db22f8db6925"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:17:27 crc kubenswrapper[4718]: I1206 04:17:27.879154 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://ab28e895cdd83b428eae8d7bd3a859b7eb41e3f8218509581d34db22f8db6925" gracePeriod=600 Dec 06 04:17:28 crc kubenswrapper[4718]: I1206 04:17:28.716514 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="ab28e895cdd83b428eae8d7bd3a859b7eb41e3f8218509581d34db22f8db6925" exitCode=0 Dec 06 04:17:28 crc kubenswrapper[4718]: I1206 04:17:28.716635 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"ab28e895cdd83b428eae8d7bd3a859b7eb41e3f8218509581d34db22f8db6925"} Dec 06 04:17:28 crc kubenswrapper[4718]: I1206 04:17:28.716812 4718 scope.go:117] "RemoveContainer" containerID="c4e3740d9b890ba2b04c60617f993f3387fea120a80ab270801f84ab026d7dd7" Dec 06 04:17:29 crc kubenswrapper[4718]: I1206 04:17:29.724754 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"5d80f8884240523b385c679ab95f41f5a9a967b7f7ed7e01c46ae96cdb397ec1"} Dec 06 04:19:30 crc kubenswrapper[4718]: I1206 04:19:30.609025 4718 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 04:19:41 crc kubenswrapper[4718]: I1206 04:19:41.755347 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6nfdl"] Dec 06 04:19:41 crc kubenswrapper[4718]: I1206 04:19:41.758743 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovn-controller" containerID="cri-o://a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884" gracePeriod=30 Dec 06 04:19:41 crc kubenswrapper[4718]: I1206 04:19:41.759744 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="sbdb" containerID="cri-o://2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" gracePeriod=30 Dec 06 04:19:41 crc kubenswrapper[4718]: I1206 04:19:41.759856 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="nbdb" containerID="cri-o://2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" gracePeriod=30 Dec 06 04:19:41 crc kubenswrapper[4718]: I1206 04:19:41.759951 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="northd" containerID="cri-o://2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5" gracePeriod=30 Dec 06 04:19:41 crc kubenswrapper[4718]: I1206 04:19:41.760036 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e" gracePeriod=30 Dec 06 04:19:41 crc kubenswrapper[4718]: I1206 04:19:41.760127 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kube-rbac-proxy-node" containerID="cri-o://a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b" gracePeriod=30 Dec 06 04:19:41 crc kubenswrapper[4718]: I1206 04:19:41.760210 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovn-acl-logging" containerID="cri-o://78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a" gracePeriod=30 Dec 06 04:19:41 crc kubenswrapper[4718]: I1206 04:19:41.812638 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" containerID="cri-o://541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5" gracePeriod=30 Dec 06 04:19:41 crc kubenswrapper[4718]: E1206 04:19:41.893931 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 06 04:19:41 crc kubenswrapper[4718]: E1206 04:19:41.894091 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 06 04:19:41 crc kubenswrapper[4718]: E1206 04:19:41.895314 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 06 04:19:41 crc kubenswrapper[4718]: E1206 04:19:41.895848 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 06 04:19:41 crc kubenswrapper[4718]: E1206 04:19:41.896634 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 06 04:19:41 crc kubenswrapper[4718]: E1206 04:19:41.896668 4718 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="sbdb" Dec 06 04:19:41 crc kubenswrapper[4718]: E1206 04:19:41.896941 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 06 04:19:41 crc kubenswrapper[4718]: E1206 04:19:41.896970 4718 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="nbdb" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.043698 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/3.log" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.046404 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovn-acl-logging/0.log" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.047047 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovn-controller/0.log" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.047682 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.094937 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-env-overrides\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.094997 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-log-socket\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095027 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-node-log\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095048 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-netns\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095089 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-kubelet\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095111 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-openvswitch\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095136 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-bin\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095158 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-systemd-units\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095183 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e9ed5db1-456b-4891-a750-809728e9c0a5-ovn-node-metrics-cert\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095211 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095277 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-ovn\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095300 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-etc-openvswitch\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095322 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-ovn-kubernetes\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095346 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-netd\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095373 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-script-lib\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095400 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnpgw\" (UniqueName: \"kubernetes.io/projected/e9ed5db1-456b-4891-a750-809728e9c0a5-kube-api-access-qnpgw\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095419 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-slash\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095476 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-systemd\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095499 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-config\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095521 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-var-lib-openvswitch\") pod \"e9ed5db1-456b-4891-a750-809728e9c0a5\" (UID: \"e9ed5db1-456b-4891-a750-809728e9c0a5\") " Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095785 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095785 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095811 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095828 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095842 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095857 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095872 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095896 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.095918 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.096015 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.096380 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.097802 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.097948 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-log-socket" (OuterVolumeSpecName: "log-socket") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.098055 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-slash" (OuterVolumeSpecName: "host-slash") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.098439 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.098818 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.099092 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-node-log" (OuterVolumeSpecName: "node-log") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.101886 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9ed5db1-456b-4891-a750-809728e9c0a5-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.102876 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9ed5db1-456b-4891-a750-809728e9c0a5-kube-api-access-qnpgw" (OuterVolumeSpecName: "kube-api-access-qnpgw") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "kube-api-access-qnpgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.109168 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "e9ed5db1-456b-4891-a750-809728e9c0a5" (UID: "e9ed5db1-456b-4891-a750-809728e9c0a5"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.110916 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bjd55"] Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115389 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="northd" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115415 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="northd" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115428 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5" containerName="collect-profiles" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115440 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5" containerName="collect-profiles" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115457 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovn-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115469 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovn-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115481 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kube-rbac-proxy-node" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115493 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kube-rbac-proxy-node" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115509 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="sbdb" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115519 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="sbdb" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115536 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115548 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115563 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kubecfg-setup" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115576 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kubecfg-setup" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115661 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="nbdb" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115678 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="nbdb" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115696 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115708 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115750 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovn-acl-logging" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115761 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovn-acl-logging" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115777 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115788 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115800 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115811 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.115821 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.115832 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116016 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovn-acl-logging" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116032 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116046 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kube-rbac-proxy-node" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116109 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116130 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="northd" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116149 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116161 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb8d8d9-1e48-4bac-a3f8-e5da23b983c5" containerName="collect-profiles" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116175 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="sbdb" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116193 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="nbdb" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116207 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116267 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovn-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.116418 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116431 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.116574 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.118795 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerName="ovnkube-controller" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.121080 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196533 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8814595d-617c-4c09-87fd-82ba454ba573-env-overrides\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196572 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-systemd-units\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196588 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-node-log\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196607 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-run-systemd\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196627 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-run-openvswitch\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196645 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-run-netns\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196661 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8814595d-617c-4c09-87fd-82ba454ba573-ovnkube-script-lib\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196678 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-etc-openvswitch\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196693 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8814595d-617c-4c09-87fd-82ba454ba573-ovn-node-metrics-cert\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196710 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pts6\" (UniqueName: \"kubernetes.io/projected/8814595d-617c-4c09-87fd-82ba454ba573-kube-api-access-6pts6\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196725 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-run-ovn\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196745 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-run-ovn-kubernetes\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196761 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-kubelet\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196774 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-cni-netd\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196790 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-cni-bin\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196805 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8814595d-617c-4c09-87fd-82ba454ba573-ovnkube-config\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196823 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196838 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-slash\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196854 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-var-lib-openvswitch\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196887 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-log-socket\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196920 4718 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196930 4718 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196938 4718 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196946 4718 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196955 4718 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e9ed5db1-456b-4891-a750-809728e9c0a5-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196963 4718 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196972 4718 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196980 4718 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196990 4718 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.196997 4718 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197005 4718 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197013 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnpgw\" (UniqueName: \"kubernetes.io/projected/e9ed5db1-456b-4891-a750-809728e9c0a5-kube-api-access-qnpgw\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197021 4718 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-slash\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197028 4718 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197036 4718 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197044 4718 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197053 4718 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e9ed5db1-456b-4891-a750-809728e9c0a5-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197061 4718 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-log-socket\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197068 4718 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-node-log\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.197076 4718 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e9ed5db1-456b-4891-a750-809728e9c0a5-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.298637 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-var-lib-openvswitch\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.298731 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-log-socket\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.298746 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-var-lib-openvswitch\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.298771 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8814595d-617c-4c09-87fd-82ba454ba573-env-overrides\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.298844 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-node-log\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.298883 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-systemd-units\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.298904 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-log-socket\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.298959 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-run-systemd\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.298924 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-run-systemd\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299005 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-node-log\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299042 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-systemd-units\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299065 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-run-openvswitch\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299109 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-run-netns\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299139 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8814595d-617c-4c09-87fd-82ba454ba573-ovnkube-script-lib\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299184 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-run-openvswitch\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299188 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-etc-openvswitch\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299226 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8814595d-617c-4c09-87fd-82ba454ba573-ovn-node-metrics-cert\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299264 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-run-netns\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299277 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pts6\" (UniqueName: \"kubernetes.io/projected/8814595d-617c-4c09-87fd-82ba454ba573-kube-api-access-6pts6\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299309 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-etc-openvswitch\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299318 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-run-ovn\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299377 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-run-ovn-kubernetes\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299404 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-kubelet\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299439 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-cni-netd\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299460 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-cni-bin\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299483 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8814595d-617c-4c09-87fd-82ba454ba573-ovnkube-config\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299523 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299552 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-slash\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299635 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-slash\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299708 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8814595d-617c-4c09-87fd-82ba454ba573-env-overrides\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.299785 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-kubelet\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.300218 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-run-ovn\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.300303 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-run-ovn-kubernetes\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.300322 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8814595d-617c-4c09-87fd-82ba454ba573-ovnkube-script-lib\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.300779 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-cni-netd\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.300882 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.301182 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8814595d-617c-4c09-87fd-82ba454ba573-ovnkube-config\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.302168 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8814595d-617c-4c09-87fd-82ba454ba573-host-cni-bin\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.306876 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8814595d-617c-4c09-87fd-82ba454ba573-ovn-node-metrics-cert\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.329185 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pts6\" (UniqueName: \"kubernetes.io/projected/8814595d-617c-4c09-87fd-82ba454ba573-kube-api-access-6pts6\") pod \"ovnkube-node-bjd55\" (UID: \"8814595d-617c-4c09-87fd-82ba454ba573\") " pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.444271 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.641120 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4r7xc_2272fbce-ce87-4f9f-b1c5-f83f262534de/kube-multus/2.log" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.641761 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4r7xc_2272fbce-ce87-4f9f-b1c5-f83f262534de/kube-multus/1.log" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.641904 4718 generic.go:334] "Generic (PLEG): container finished" podID="2272fbce-ce87-4f9f-b1c5-f83f262534de" containerID="5dc497939199c4b153bf5866b448171883409c3dfde61d01d920b1194fc12c52" exitCode=2 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.641999 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4r7xc" event={"ID":"2272fbce-ce87-4f9f-b1c5-f83f262534de","Type":"ContainerDied","Data":"5dc497939199c4b153bf5866b448171883409c3dfde61d01d920b1194fc12c52"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.642192 4718 scope.go:117] "RemoveContainer" containerID="c0f98ab4781ffee60af4d1d9be5b040118fb062bbf03fc1ec0013e519fe0f31a" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.642658 4718 scope.go:117] "RemoveContainer" containerID="5dc497939199c4b153bf5866b448171883409c3dfde61d01d920b1194fc12c52" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.644484 4718 generic.go:334] "Generic (PLEG): container finished" podID="8814595d-617c-4c09-87fd-82ba454ba573" containerID="392d3f5ab089efe5c9efabfd849a688e546cac4881997e6406ca013c37316097" exitCode=0 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.644616 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerDied","Data":"392d3f5ab089efe5c9efabfd849a688e546cac4881997e6406ca013c37316097"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.644662 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerStarted","Data":"7cdcd74b7e86d3680ede1c15f6972667db8ec985cfddcbb0852cc64a18f8876b"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.649487 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovnkube-controller/3.log" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.653542 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovn-acl-logging/0.log" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654315 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6nfdl_e9ed5db1-456b-4891-a750-809728e9c0a5/ovn-controller/0.log" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654741 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5" exitCode=0 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654780 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" exitCode=0 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654797 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" exitCode=0 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654844 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5" exitCode=0 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654861 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e" exitCode=0 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654876 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b" exitCode=0 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654889 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a" exitCode=143 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654902 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9ed5db1-456b-4891-a750-809728e9c0a5" containerID="a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884" exitCode=143 Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654925 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.654912 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655023 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655043 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655440 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655515 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655536 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655553 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655570 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655581 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655591 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655600 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655610 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655619 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655629 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655638 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655648 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655661 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655678 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655689 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655699 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655709 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655718 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655727 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655739 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655749 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655758 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655767 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655780 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655795 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655807 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655817 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655827 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655836 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655846 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655856 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655865 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655874 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655886 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655899 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6nfdl" event={"ID":"e9ed5db1-456b-4891-a750-809728e9c0a5","Type":"ContainerDied","Data":"1a7fb23f2bc1e4d7ea3c66d92ddb9fc20b39c4f38194f736af44dfb9315241bc"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.655914 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.657405 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.657498 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.657574 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.657639 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.657700 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.657769 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.657838 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.657917 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.658000 4718 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe"} Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.683463 4718 scope.go:117] "RemoveContainer" containerID="541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.735796 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6nfdl"] Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.737385 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.738390 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6nfdl"] Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.760414 4718 scope.go:117] "RemoveContainer" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.777053 4718 scope.go:117] "RemoveContainer" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.791571 4718 scope.go:117] "RemoveContainer" containerID="2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.808773 4718 scope.go:117] "RemoveContainer" containerID="16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.834313 4718 scope.go:117] "RemoveContainer" containerID="a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.848083 4718 scope.go:117] "RemoveContainer" containerID="78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.865543 4718 scope.go:117] "RemoveContainer" containerID="a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.896100 4718 scope.go:117] "RemoveContainer" containerID="81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.919795 4718 scope.go:117] "RemoveContainer" containerID="541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.920402 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5\": container with ID starting with 541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5 not found: ID does not exist" containerID="541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.920439 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} err="failed to get container status \"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5\": rpc error: code = NotFound desc = could not find container \"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5\": container with ID starting with 541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.920467 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.920809 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\": container with ID starting with 6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de not found: ID does not exist" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.920843 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} err="failed to get container status \"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\": rpc error: code = NotFound desc = could not find container \"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\": container with ID starting with 6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.920861 4718 scope.go:117] "RemoveContainer" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.921303 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\": container with ID starting with 2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35 not found: ID does not exist" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.921333 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} err="failed to get container status \"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\": rpc error: code = NotFound desc = could not find container \"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\": container with ID starting with 2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.921353 4718 scope.go:117] "RemoveContainer" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.921665 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\": container with ID starting with 2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366 not found: ID does not exist" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.921705 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} err="failed to get container status \"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\": rpc error: code = NotFound desc = could not find container \"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\": container with ID starting with 2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.921731 4718 scope.go:117] "RemoveContainer" containerID="2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.922022 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\": container with ID starting with 2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5 not found: ID does not exist" containerID="2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.922053 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} err="failed to get container status \"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\": rpc error: code = NotFound desc = could not find container \"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\": container with ID starting with 2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.922075 4718 scope.go:117] "RemoveContainer" containerID="16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.924531 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\": container with ID starting with 16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e not found: ID does not exist" containerID="16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.924558 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} err="failed to get container status \"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\": rpc error: code = NotFound desc = could not find container \"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\": container with ID starting with 16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.924576 4718 scope.go:117] "RemoveContainer" containerID="a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.924962 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\": container with ID starting with a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b not found: ID does not exist" containerID="a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.924999 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} err="failed to get container status \"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\": rpc error: code = NotFound desc = could not find container \"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\": container with ID starting with a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.925048 4718 scope.go:117] "RemoveContainer" containerID="78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.925406 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\": container with ID starting with 78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a not found: ID does not exist" containerID="78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.925428 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} err="failed to get container status \"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\": rpc error: code = NotFound desc = could not find container \"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\": container with ID starting with 78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.925442 4718 scope.go:117] "RemoveContainer" containerID="a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.925870 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\": container with ID starting with a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884 not found: ID does not exist" containerID="a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.925895 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} err="failed to get container status \"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\": rpc error: code = NotFound desc = could not find container \"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\": container with ID starting with a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.925911 4718 scope.go:117] "RemoveContainer" containerID="81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe" Dec 06 04:19:42 crc kubenswrapper[4718]: E1206 04:19:42.926817 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\": container with ID starting with 81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe not found: ID does not exist" containerID="81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.926852 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe"} err="failed to get container status \"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\": rpc error: code = NotFound desc = could not find container \"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\": container with ID starting with 81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.926884 4718 scope.go:117] "RemoveContainer" containerID="541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.927836 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} err="failed to get container status \"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5\": rpc error: code = NotFound desc = could not find container \"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5\": container with ID starting with 541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.927859 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.928633 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} err="failed to get container status \"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\": rpc error: code = NotFound desc = could not find container \"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\": container with ID starting with 6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.928664 4718 scope.go:117] "RemoveContainer" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.928963 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} err="failed to get container status \"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\": rpc error: code = NotFound desc = could not find container \"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\": container with ID starting with 2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.928995 4718 scope.go:117] "RemoveContainer" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.929382 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} err="failed to get container status \"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\": rpc error: code = NotFound desc = could not find container \"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\": container with ID starting with 2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.929414 4718 scope.go:117] "RemoveContainer" containerID="2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.929725 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} err="failed to get container status \"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\": rpc error: code = NotFound desc = could not find container \"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\": container with ID starting with 2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.929762 4718 scope.go:117] "RemoveContainer" containerID="16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.930066 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} err="failed to get container status \"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\": rpc error: code = NotFound desc = could not find container \"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\": container with ID starting with 16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.930093 4718 scope.go:117] "RemoveContainer" containerID="a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.931137 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} err="failed to get container status \"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\": rpc error: code = NotFound desc = could not find container \"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\": container with ID starting with a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.931193 4718 scope.go:117] "RemoveContainer" containerID="78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.931565 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} err="failed to get container status \"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\": rpc error: code = NotFound desc = could not find container \"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\": container with ID starting with 78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.931590 4718 scope.go:117] "RemoveContainer" containerID="a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.931895 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} err="failed to get container status \"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\": rpc error: code = NotFound desc = could not find container \"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\": container with ID starting with a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.931929 4718 scope.go:117] "RemoveContainer" containerID="81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.932199 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe"} err="failed to get container status \"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\": rpc error: code = NotFound desc = could not find container \"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\": container with ID starting with 81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.932225 4718 scope.go:117] "RemoveContainer" containerID="541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.932514 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} err="failed to get container status \"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5\": rpc error: code = NotFound desc = could not find container \"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5\": container with ID starting with 541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.932538 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.932802 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} err="failed to get container status \"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\": rpc error: code = NotFound desc = could not find container \"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\": container with ID starting with 6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.932824 4718 scope.go:117] "RemoveContainer" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.933129 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} err="failed to get container status \"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\": rpc error: code = NotFound desc = could not find container \"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\": container with ID starting with 2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.933186 4718 scope.go:117] "RemoveContainer" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.938752 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} err="failed to get container status \"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\": rpc error: code = NotFound desc = could not find container \"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\": container with ID starting with 2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.938794 4718 scope.go:117] "RemoveContainer" containerID="2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.939222 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} err="failed to get container status \"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\": rpc error: code = NotFound desc = could not find container \"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\": container with ID starting with 2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.939299 4718 scope.go:117] "RemoveContainer" containerID="16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.941400 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} err="failed to get container status \"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\": rpc error: code = NotFound desc = could not find container \"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\": container with ID starting with 16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.941427 4718 scope.go:117] "RemoveContainer" containerID="a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.941769 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} err="failed to get container status \"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\": rpc error: code = NotFound desc = could not find container \"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\": container with ID starting with a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.941794 4718 scope.go:117] "RemoveContainer" containerID="78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.942044 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} err="failed to get container status \"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\": rpc error: code = NotFound desc = could not find container \"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\": container with ID starting with 78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.942068 4718 scope.go:117] "RemoveContainer" containerID="a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.942419 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} err="failed to get container status \"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\": rpc error: code = NotFound desc = could not find container \"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\": container with ID starting with a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.942477 4718 scope.go:117] "RemoveContainer" containerID="81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.942795 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe"} err="failed to get container status \"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\": rpc error: code = NotFound desc = could not find container \"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\": container with ID starting with 81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.942820 4718 scope.go:117] "RemoveContainer" containerID="541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943056 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5"} err="failed to get container status \"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5\": rpc error: code = NotFound desc = could not find container \"541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5\": container with ID starting with 541913c6c177e39a6ff1f3bd53d1a45512660dd7f9fce0cdb753be23e4e0dcf5 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943076 4718 scope.go:117] "RemoveContainer" containerID="6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943289 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de"} err="failed to get container status \"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\": rpc error: code = NotFound desc = could not find container \"6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de\": container with ID starting with 6dc15d85f053589dc0528e8586cfed48e37373c75ff55d091a0a11867d6252de not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943310 4718 scope.go:117] "RemoveContainer" containerID="2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943490 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35"} err="failed to get container status \"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\": rpc error: code = NotFound desc = could not find container \"2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35\": container with ID starting with 2f1e183b8ca5f3a59eeea953d494c2efd21b9b2db85969880d7b44f315cc0f35 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943514 4718 scope.go:117] "RemoveContainer" containerID="2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943682 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366"} err="failed to get container status \"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\": rpc error: code = NotFound desc = could not find container \"2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366\": container with ID starting with 2436704b97e9111315ed3a8ec76b23970e94f19ba4ad75ee561cf48a30201366 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943706 4718 scope.go:117] "RemoveContainer" containerID="2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943872 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5"} err="failed to get container status \"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\": rpc error: code = NotFound desc = could not find container \"2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5\": container with ID starting with 2b1771a48f5d916522dcb829a9ca6809ec436ef050e07c001420b06085ea7ac5 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.943891 4718 scope.go:117] "RemoveContainer" containerID="16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.944058 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e"} err="failed to get container status \"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\": rpc error: code = NotFound desc = could not find container \"16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e\": container with ID starting with 16bee2d39752654b8fc07c7422f2b40eb876022712a43998f401c6cd780f7e6e not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.944079 4718 scope.go:117] "RemoveContainer" containerID="a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.944257 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b"} err="failed to get container status \"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\": rpc error: code = NotFound desc = could not find container \"a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b\": container with ID starting with a83ff8294496c557e5a5aa0e60bc8889113267e0f54a134e0898b5536c43649b not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.944275 4718 scope.go:117] "RemoveContainer" containerID="78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.944499 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a"} err="failed to get container status \"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\": rpc error: code = NotFound desc = could not find container \"78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a\": container with ID starting with 78e09a8a20838f9175a3562ecbce54021327796ca4066540b733eb6a1292f37a not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.944537 4718 scope.go:117] "RemoveContainer" containerID="a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.944767 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884"} err="failed to get container status \"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\": rpc error: code = NotFound desc = could not find container \"a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884\": container with ID starting with a54eae4437ba63601b145244e7a6d8f92cfaa25500e2f90a41e1e0030d5d7884 not found: ID does not exist" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.944793 4718 scope.go:117] "RemoveContainer" containerID="81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe" Dec 06 04:19:42 crc kubenswrapper[4718]: I1206 04:19:42.944971 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe"} err="failed to get container status \"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\": rpc error: code = NotFound desc = could not find container \"81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe\": container with ID starting with 81585ef68d7d66e45637ca5881ecb2a6e524aa2c0cd0b9f256bbc037179876fe not found: ID does not exist" Dec 06 04:19:43 crc kubenswrapper[4718]: I1206 04:19:43.337136 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9ed5db1-456b-4891-a750-809728e9c0a5" path="/var/lib/kubelet/pods/e9ed5db1-456b-4891-a750-809728e9c0a5/volumes" Dec 06 04:19:43 crc kubenswrapper[4718]: I1206 04:19:43.662304 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4r7xc_2272fbce-ce87-4f9f-b1c5-f83f262534de/kube-multus/2.log" Dec 06 04:19:43 crc kubenswrapper[4718]: I1206 04:19:43.662636 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4r7xc" event={"ID":"2272fbce-ce87-4f9f-b1c5-f83f262534de","Type":"ContainerStarted","Data":"79f3a6e56367689b1b3392fd56c8133ef9636bd0bd04f413253518f846c1f054"} Dec 06 04:19:43 crc kubenswrapper[4718]: I1206 04:19:43.667751 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerStarted","Data":"2bb5c25c2a62db9ef080ab30df7e5e20d5bce24e7ba9cc29d8d167658a420059"} Dec 06 04:19:43 crc kubenswrapper[4718]: I1206 04:19:43.667943 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerStarted","Data":"dbe4dbcff26462571242e4fb36ec0d803d73c361c18ab42a5fed9c0028c54127"} Dec 06 04:19:43 crc kubenswrapper[4718]: I1206 04:19:43.668045 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerStarted","Data":"d22b1353bf7a054fc90c724094f3d3a286475802f6e22329b279eb149c0ffade"} Dec 06 04:19:43 crc kubenswrapper[4718]: I1206 04:19:43.668123 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerStarted","Data":"4c9717156ba5bd9532697959c0ad7e281e868e0d1c40082b70afbbffadfd9ce3"} Dec 06 04:19:43 crc kubenswrapper[4718]: I1206 04:19:43.668192 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerStarted","Data":"09ac24a0d96f46134f26ba1e138ac0c1d50b2b4b6abd8cd00adbc76a397db359"} Dec 06 04:19:43 crc kubenswrapper[4718]: I1206 04:19:43.668306 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerStarted","Data":"72da42432008bd73ca5c0c48fe2bad2b24eb662a005656349348056de74b985e"} Dec 06 04:19:45 crc kubenswrapper[4718]: I1206 04:19:45.690783 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerStarted","Data":"99f36f28b233c58e610e8e598d19482c9352efee5468b71e52a980efa091fbe6"} Dec 06 04:19:48 crc kubenswrapper[4718]: I1206 04:19:48.713209 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" event={"ID":"8814595d-617c-4c09-87fd-82ba454ba573","Type":"ContainerStarted","Data":"5e394a124448672f3724c009bd393e49fea82c55e2970df8cc9d0473a09767d9"} Dec 06 04:19:48 crc kubenswrapper[4718]: I1206 04:19:48.714268 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:48 crc kubenswrapper[4718]: I1206 04:19:48.748357 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" podStartSLOduration=6.74833299 podStartE2EDuration="6.74833299s" podCreationTimestamp="2025-12-06 04:19:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:19:48.740841637 +0000 UTC m=+777.746546838" watchObservedRunningTime="2025-12-06 04:19:48.74833299 +0000 UTC m=+777.754038191" Dec 06 04:19:48 crc kubenswrapper[4718]: I1206 04:19:48.749526 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:49 crc kubenswrapper[4718]: I1206 04:19:49.721298 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:49 crc kubenswrapper[4718]: I1206 04:19:49.721753 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:49 crc kubenswrapper[4718]: I1206 04:19:49.761005 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:19:57 crc kubenswrapper[4718]: I1206 04:19:57.878178 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:19:57 crc kubenswrapper[4718]: I1206 04:19:57.879210 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:20:10 crc kubenswrapper[4718]: I1206 04:20:10.717462 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-zgxbt"] Dec 06 04:20:10 crc kubenswrapper[4718]: I1206 04:20:10.719016 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:20:10 crc kubenswrapper[4718]: I1206 04:20:10.721506 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-dgzkg" Dec 06 04:20:10 crc kubenswrapper[4718]: I1206 04:20:10.724691 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 06 04:20:10 crc kubenswrapper[4718]: I1206 04:20:10.724984 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 06 04:20:10 crc kubenswrapper[4718]: I1206 04:20:10.787737 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-zgxbt"] Dec 06 04:20:10 crc kubenswrapper[4718]: I1206 04:20:10.830455 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9qwn\" (UniqueName: \"kubernetes.io/projected/f6373f59-2255-4710-b1c2-e2d2c4624289-kube-api-access-h9qwn\") pod \"mariadb-operator-index-zgxbt\" (UID: \"f6373f59-2255-4710-b1c2-e2d2c4624289\") " pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:20:10 crc kubenswrapper[4718]: I1206 04:20:10.932193 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9qwn\" (UniqueName: \"kubernetes.io/projected/f6373f59-2255-4710-b1c2-e2d2c4624289-kube-api-access-h9qwn\") pod \"mariadb-operator-index-zgxbt\" (UID: \"f6373f59-2255-4710-b1c2-e2d2c4624289\") " pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:20:10 crc kubenswrapper[4718]: I1206 04:20:10.954820 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9qwn\" (UniqueName: \"kubernetes.io/projected/f6373f59-2255-4710-b1c2-e2d2c4624289-kube-api-access-h9qwn\") pod \"mariadb-operator-index-zgxbt\" (UID: \"f6373f59-2255-4710-b1c2-e2d2c4624289\") " pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:20:11 crc kubenswrapper[4718]: I1206 04:20:11.097644 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:20:11 crc kubenswrapper[4718]: I1206 04:20:11.558042 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-zgxbt"] Dec 06 04:20:11 crc kubenswrapper[4718]: I1206 04:20:11.564797 4718 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:20:11 crc kubenswrapper[4718]: I1206 04:20:11.881904 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zgxbt" event={"ID":"f6373f59-2255-4710-b1c2-e2d2c4624289","Type":"ContainerStarted","Data":"8b2089920bb2df6a4b57fde423ed2015d92cd20bf81390326164bd61c9bec54c"} Dec 06 04:20:12 crc kubenswrapper[4718]: I1206 04:20:12.468746 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bjd55" Dec 06 04:20:12 crc kubenswrapper[4718]: I1206 04:20:12.889585 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zgxbt" event={"ID":"f6373f59-2255-4710-b1c2-e2d2c4624289","Type":"ContainerStarted","Data":"175b20f6fd66204f9be95f80d7dae08ed830a2f29f7eb7dccffed6ead1bd7c54"} Dec 06 04:20:12 crc kubenswrapper[4718]: I1206 04:20:12.910423 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-zgxbt" podStartSLOduration=2.153809125 podStartE2EDuration="2.910396058s" podCreationTimestamp="2025-12-06 04:20:10 +0000 UTC" firstStartedPulling="2025-12-06 04:20:11.564549831 +0000 UTC m=+800.570254992" lastFinishedPulling="2025-12-06 04:20:12.321136774 +0000 UTC m=+801.326841925" observedRunningTime="2025-12-06 04:20:12.904204723 +0000 UTC m=+801.909909884" watchObservedRunningTime="2025-12-06 04:20:12.910396058 +0000 UTC m=+801.916101249" Dec 06 04:20:21 crc kubenswrapper[4718]: I1206 04:20:21.098550 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:20:21 crc kubenswrapper[4718]: I1206 04:20:21.099367 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:20:21 crc kubenswrapper[4718]: I1206 04:20:21.139984 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:20:21 crc kubenswrapper[4718]: I1206 04:20:21.987111 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.721074 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx"] Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.723801 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.729726 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-69frt" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.739500 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx"] Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.877747 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.877821 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.888790 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-util\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.888852 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-bundle\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.888890 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbld7\" (UniqueName: \"kubernetes.io/projected/e7325a25-f02a-4afa-867f-98255311c258-kube-api-access-hbld7\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.989903 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-util\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.989973 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-bundle\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.990023 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbld7\" (UniqueName: \"kubernetes.io/projected/e7325a25-f02a-4afa-867f-98255311c258-kube-api-access-hbld7\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.990569 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-util\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:27 crc kubenswrapper[4718]: I1206 04:20:27.990609 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-bundle\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:28 crc kubenswrapper[4718]: I1206 04:20:28.013342 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbld7\" (UniqueName: \"kubernetes.io/projected/e7325a25-f02a-4afa-867f-98255311c258-kube-api-access-hbld7\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:28 crc kubenswrapper[4718]: I1206 04:20:28.085821 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:28 crc kubenswrapper[4718]: I1206 04:20:28.263835 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx"] Dec 06 04:20:28 crc kubenswrapper[4718]: I1206 04:20:28.995038 4718 generic.go:334] "Generic (PLEG): container finished" podID="e7325a25-f02a-4afa-867f-98255311c258" containerID="8a56447d0b735769e654f4b328cc14bd65ec17aa99f7bdb216d0fb0d427661fd" exitCode=0 Dec 06 04:20:28 crc kubenswrapper[4718]: I1206 04:20:28.995107 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" event={"ID":"e7325a25-f02a-4afa-867f-98255311c258","Type":"ContainerDied","Data":"8a56447d0b735769e654f4b328cc14bd65ec17aa99f7bdb216d0fb0d427661fd"} Dec 06 04:20:28 crc kubenswrapper[4718]: I1206 04:20:28.995149 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" event={"ID":"e7325a25-f02a-4afa-867f-98255311c258","Type":"ContainerStarted","Data":"cfb9cd8aa5b57dc547443b76a0e6825a99ce457d8af922037f737dfd4c1736fe"} Dec 06 04:20:31 crc kubenswrapper[4718]: I1206 04:20:31.054547 4718 generic.go:334] "Generic (PLEG): container finished" podID="e7325a25-f02a-4afa-867f-98255311c258" containerID="2d1388cbab7597dc1d9bd2d00889495b9b31924c1932ca8836a9d9ccf35076f8" exitCode=0 Dec 06 04:20:31 crc kubenswrapper[4718]: I1206 04:20:31.054737 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" event={"ID":"e7325a25-f02a-4afa-867f-98255311c258","Type":"ContainerDied","Data":"2d1388cbab7597dc1d9bd2d00889495b9b31924c1932ca8836a9d9ccf35076f8"} Dec 06 04:20:32 crc kubenswrapper[4718]: I1206 04:20:32.065603 4718 generic.go:334] "Generic (PLEG): container finished" podID="e7325a25-f02a-4afa-867f-98255311c258" containerID="3f8b90b980ff5b05fdca17e34d36049ee00b9960b8169dfe6a5c32b0e52bc034" exitCode=0 Dec 06 04:20:32 crc kubenswrapper[4718]: I1206 04:20:32.065787 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" event={"ID":"e7325a25-f02a-4afa-867f-98255311c258","Type":"ContainerDied","Data":"3f8b90b980ff5b05fdca17e34d36049ee00b9960b8169dfe6a5c32b0e52bc034"} Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.414879 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.587109 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-util\") pod \"e7325a25-f02a-4afa-867f-98255311c258\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.587174 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbld7\" (UniqueName: \"kubernetes.io/projected/e7325a25-f02a-4afa-867f-98255311c258-kube-api-access-hbld7\") pod \"e7325a25-f02a-4afa-867f-98255311c258\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.587248 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-bundle\") pod \"e7325a25-f02a-4afa-867f-98255311c258\" (UID: \"e7325a25-f02a-4afa-867f-98255311c258\") " Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.588600 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-bundle" (OuterVolumeSpecName: "bundle") pod "e7325a25-f02a-4afa-867f-98255311c258" (UID: "e7325a25-f02a-4afa-867f-98255311c258"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.593346 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7325a25-f02a-4afa-867f-98255311c258-kube-api-access-hbld7" (OuterVolumeSpecName: "kube-api-access-hbld7") pod "e7325a25-f02a-4afa-867f-98255311c258" (UID: "e7325a25-f02a-4afa-867f-98255311c258"). InnerVolumeSpecName "kube-api-access-hbld7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.605371 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-util" (OuterVolumeSpecName: "util") pod "e7325a25-f02a-4afa-867f-98255311c258" (UID: "e7325a25-f02a-4afa-867f-98255311c258"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.689018 4718 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-util\") on node \"crc\" DevicePath \"\"" Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.689059 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbld7\" (UniqueName: \"kubernetes.io/projected/e7325a25-f02a-4afa-867f-98255311c258-kube-api-access-hbld7\") on node \"crc\" DevicePath \"\"" Dec 06 04:20:33 crc kubenswrapper[4718]: I1206 04:20:33.689075 4718 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7325a25-f02a-4afa-867f-98255311c258-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:20:34 crc kubenswrapper[4718]: I1206 04:20:34.080822 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" event={"ID":"e7325a25-f02a-4afa-867f-98255311c258","Type":"ContainerDied","Data":"cfb9cd8aa5b57dc547443b76a0e6825a99ce457d8af922037f737dfd4c1736fe"} Dec 06 04:20:34 crc kubenswrapper[4718]: I1206 04:20:34.080859 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfb9cd8aa5b57dc547443b76a0e6825a99ce457d8af922037f737dfd4c1736fe" Dec 06 04:20:34 crc kubenswrapper[4718]: I1206 04:20:34.080900 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.015905 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph"] Dec 06 04:20:41 crc kubenswrapper[4718]: E1206 04:20:41.016835 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7325a25-f02a-4afa-867f-98255311c258" containerName="pull" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.016854 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7325a25-f02a-4afa-867f-98255311c258" containerName="pull" Dec 06 04:20:41 crc kubenswrapper[4718]: E1206 04:20:41.016876 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7325a25-f02a-4afa-867f-98255311c258" containerName="extract" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.016889 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7325a25-f02a-4afa-867f-98255311c258" containerName="extract" Dec 06 04:20:41 crc kubenswrapper[4718]: E1206 04:20:41.016913 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7325a25-f02a-4afa-867f-98255311c258" containerName="util" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.016922 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7325a25-f02a-4afa-867f-98255311c258" containerName="util" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.017064 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7325a25-f02a-4afa-867f-98255311c258" containerName="extract" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.017628 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.020603 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.021365 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.021985 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-wl5xc" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.050199 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph"] Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.082046 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w89n\" (UniqueName: \"kubernetes.io/projected/043bf4f1-afde-4885-ab7a-764c626964c2-kube-api-access-2w89n\") pod \"mariadb-operator-controller-manager-548f66d7bb-wn4ph\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.082116 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-apiservice-cert\") pod \"mariadb-operator-controller-manager-548f66d7bb-wn4ph\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.082191 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-webhook-cert\") pod \"mariadb-operator-controller-manager-548f66d7bb-wn4ph\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.182625 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w89n\" (UniqueName: \"kubernetes.io/projected/043bf4f1-afde-4885-ab7a-764c626964c2-kube-api-access-2w89n\") pod \"mariadb-operator-controller-manager-548f66d7bb-wn4ph\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.182917 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-apiservice-cert\") pod \"mariadb-operator-controller-manager-548f66d7bb-wn4ph\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.182968 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-webhook-cert\") pod \"mariadb-operator-controller-manager-548f66d7bb-wn4ph\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.193278 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-webhook-cert\") pod \"mariadb-operator-controller-manager-548f66d7bb-wn4ph\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.198754 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-apiservice-cert\") pod \"mariadb-operator-controller-manager-548f66d7bb-wn4ph\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.203170 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w89n\" (UniqueName: \"kubernetes.io/projected/043bf4f1-afde-4885-ab7a-764c626964c2-kube-api-access-2w89n\") pod \"mariadb-operator-controller-manager-548f66d7bb-wn4ph\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.347212 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:41 crc kubenswrapper[4718]: I1206 04:20:41.595962 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph"] Dec 06 04:20:42 crc kubenswrapper[4718]: I1206 04:20:42.126096 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" event={"ID":"043bf4f1-afde-4885-ab7a-764c626964c2","Type":"ContainerStarted","Data":"39b2b1fd2ec436722773fa99932d4c6d2ac400ddbf935c56f4409652d529795b"} Dec 06 04:20:45 crc kubenswrapper[4718]: I1206 04:20:45.141516 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" event={"ID":"043bf4f1-afde-4885-ab7a-764c626964c2","Type":"ContainerStarted","Data":"101fe35aa482af5d13bc0b558f367dad441240baacbb4b3557992a9703e4306c"} Dec 06 04:20:45 crc kubenswrapper[4718]: I1206 04:20:45.142147 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:45 crc kubenswrapper[4718]: I1206 04:20:45.160565 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" podStartSLOduration=2.04715385 podStartE2EDuration="5.160550843s" podCreationTimestamp="2025-12-06 04:20:40 +0000 UTC" firstStartedPulling="2025-12-06 04:20:41.604434845 +0000 UTC m=+830.610140016" lastFinishedPulling="2025-12-06 04:20:44.717831848 +0000 UTC m=+833.723537009" observedRunningTime="2025-12-06 04:20:45.157147836 +0000 UTC m=+834.162853067" watchObservedRunningTime="2025-12-06 04:20:45.160550843 +0000 UTC m=+834.166256004" Dec 06 04:20:51 crc kubenswrapper[4718]: I1206 04:20:51.350847 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:20:52 crc kubenswrapper[4718]: I1206 04:20:52.918959 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7"] Dec 06 04:20:52 crc kubenswrapper[4718]: I1206 04:20:52.920549 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:52 crc kubenswrapper[4718]: I1206 04:20:52.922879 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 04:20:52 crc kubenswrapper[4718]: I1206 04:20:52.929053 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7"] Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.058314 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb4lj\" (UniqueName: \"kubernetes.io/projected/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-kube-api-access-fb4lj\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.058372 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.058452 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.159183 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb4lj\" (UniqueName: \"kubernetes.io/projected/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-kube-api-access-fb4lj\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.159266 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.159301 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.159778 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.159852 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.191015 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb4lj\" (UniqueName: \"kubernetes.io/projected/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-kube-api-access-fb4lj\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.243138 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:53 crc kubenswrapper[4718]: I1206 04:20:53.805572 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7"] Dec 06 04:20:53 crc kubenswrapper[4718]: W1206 04:20:53.821677 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf08e6b54_9291_41eb_9b5c_4265e8e25eb9.slice/crio-38f6eec76c1b7b20e2828c671196847a669acaa15210467c3be11ee7c17a963f WatchSource:0}: Error finding container 38f6eec76c1b7b20e2828c671196847a669acaa15210467c3be11ee7c17a963f: Status 404 returned error can't find the container with id 38f6eec76c1b7b20e2828c671196847a669acaa15210467c3be11ee7c17a963f Dec 06 04:20:54 crc kubenswrapper[4718]: I1206 04:20:54.192876 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" event={"ID":"f08e6b54-9291-41eb-9b5c-4265e8e25eb9","Type":"ContainerStarted","Data":"764c7021ad2b990c08c076e8331d8f751512a34cc3f2195d24993c7d809f7242"} Dec 06 04:20:54 crc kubenswrapper[4718]: I1206 04:20:54.192919 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" event={"ID":"f08e6b54-9291-41eb-9b5c-4265e8e25eb9","Type":"ContainerStarted","Data":"38f6eec76c1b7b20e2828c671196847a669acaa15210467c3be11ee7c17a963f"} Dec 06 04:20:55 crc kubenswrapper[4718]: I1206 04:20:55.201161 4718 generic.go:334] "Generic (PLEG): container finished" podID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerID="764c7021ad2b990c08c076e8331d8f751512a34cc3f2195d24993c7d809f7242" exitCode=0 Dec 06 04:20:55 crc kubenswrapper[4718]: I1206 04:20:55.201282 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" event={"ID":"f08e6b54-9291-41eb-9b5c-4265e8e25eb9","Type":"ContainerDied","Data":"764c7021ad2b990c08c076e8331d8f751512a34cc3f2195d24993c7d809f7242"} Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.472435 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4z8ml"] Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.474974 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.491856 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4z8ml"] Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.612365 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-catalog-content\") pod \"redhat-operators-4z8ml\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.612795 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjvrb\" (UniqueName: \"kubernetes.io/projected/155c38ca-0913-47f0-9cb1-21063447f752-kube-api-access-xjvrb\") pod \"redhat-operators-4z8ml\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.612909 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-utilities\") pod \"redhat-operators-4z8ml\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.714001 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-catalog-content\") pod \"redhat-operators-4z8ml\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.714105 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjvrb\" (UniqueName: \"kubernetes.io/projected/155c38ca-0913-47f0-9cb1-21063447f752-kube-api-access-xjvrb\") pod \"redhat-operators-4z8ml\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.714133 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-utilities\") pod \"redhat-operators-4z8ml\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.714611 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-catalog-content\") pod \"redhat-operators-4z8ml\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.714644 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-utilities\") pod \"redhat-operators-4z8ml\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.736646 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjvrb\" (UniqueName: \"kubernetes.io/projected/155c38ca-0913-47f0-9cb1-21063447f752-kube-api-access-xjvrb\") pod \"redhat-operators-4z8ml\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:56 crc kubenswrapper[4718]: I1206 04:20:56.802513 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.011830 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4z8ml"] Dec 06 04:20:57 crc kubenswrapper[4718]: W1206 04:20:57.020776 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod155c38ca_0913_47f0_9cb1_21063447f752.slice/crio-b7de719cb3549dbcc90d9cd347964e6eaa90f18fdeb9df3121fe339483faaaf2 WatchSource:0}: Error finding container b7de719cb3549dbcc90d9cd347964e6eaa90f18fdeb9df3121fe339483faaaf2: Status 404 returned error can't find the container with id b7de719cb3549dbcc90d9cd347964e6eaa90f18fdeb9df3121fe339483faaaf2 Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.214921 4718 generic.go:334] "Generic (PLEG): container finished" podID="155c38ca-0913-47f0-9cb1-21063447f752" containerID="555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6" exitCode=0 Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.215018 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4z8ml" event={"ID":"155c38ca-0913-47f0-9cb1-21063447f752","Type":"ContainerDied","Data":"555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6"} Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.215304 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4z8ml" event={"ID":"155c38ca-0913-47f0-9cb1-21063447f752","Type":"ContainerStarted","Data":"b7de719cb3549dbcc90d9cd347964e6eaa90f18fdeb9df3121fe339483faaaf2"} Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.217556 4718 generic.go:334] "Generic (PLEG): container finished" podID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerID="45fd04e497bc7837eaf3659ee8f03f4b35cf3b703fce957f400b22372a5634be" exitCode=0 Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.217593 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" event={"ID":"f08e6b54-9291-41eb-9b5c-4265e8e25eb9","Type":"ContainerDied","Data":"45fd04e497bc7837eaf3659ee8f03f4b35cf3b703fce957f400b22372a5634be"} Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.877245 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.877310 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.877368 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.877968 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d80f8884240523b385c679ab95f41f5a9a967b7f7ed7e01c46ae96cdb397ec1"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:20:57 crc kubenswrapper[4718]: I1206 04:20:57.878031 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://5d80f8884240523b385c679ab95f41f5a9a967b7f7ed7e01c46ae96cdb397ec1" gracePeriod=600 Dec 06 04:20:58 crc kubenswrapper[4718]: I1206 04:20:58.225528 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4z8ml" event={"ID":"155c38ca-0913-47f0-9cb1-21063447f752","Type":"ContainerStarted","Data":"fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f"} Dec 06 04:20:58 crc kubenswrapper[4718]: I1206 04:20:58.228969 4718 generic.go:334] "Generic (PLEG): container finished" podID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerID="089205b4773506c967e94de79adbc83fe29941024aae29cdff146e595fc7206a" exitCode=0 Dec 06 04:20:58 crc kubenswrapper[4718]: I1206 04:20:58.229112 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" event={"ID":"f08e6b54-9291-41eb-9b5c-4265e8e25eb9","Type":"ContainerDied","Data":"089205b4773506c967e94de79adbc83fe29941024aae29cdff146e595fc7206a"} Dec 06 04:20:58 crc kubenswrapper[4718]: I1206 04:20:58.231869 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="5d80f8884240523b385c679ab95f41f5a9a967b7f7ed7e01c46ae96cdb397ec1" exitCode=0 Dec 06 04:20:58 crc kubenswrapper[4718]: I1206 04:20:58.231899 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"5d80f8884240523b385c679ab95f41f5a9a967b7f7ed7e01c46ae96cdb397ec1"} Dec 06 04:20:58 crc kubenswrapper[4718]: I1206 04:20:58.231916 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"5492a0b85bf17b4550c5ff3ce07a61828c2b6f5eb097631d3cdfc69dc9a69c85"} Dec 06 04:20:58 crc kubenswrapper[4718]: I1206 04:20:58.231932 4718 scope.go:117] "RemoveContainer" containerID="ab28e895cdd83b428eae8d7bd3a859b7eb41e3f8218509581d34db22f8db6925" Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.244395 4718 generic.go:334] "Generic (PLEG): container finished" podID="155c38ca-0913-47f0-9cb1-21063447f752" containerID="fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f" exitCode=0 Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.244512 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4z8ml" event={"ID":"155c38ca-0913-47f0-9cb1-21063447f752","Type":"ContainerDied","Data":"fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f"} Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.532316 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.647791 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb4lj\" (UniqueName: \"kubernetes.io/projected/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-kube-api-access-fb4lj\") pod \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.647846 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-util\") pod \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.647918 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-bundle\") pod \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\" (UID: \"f08e6b54-9291-41eb-9b5c-4265e8e25eb9\") " Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.649703 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-bundle" (OuterVolumeSpecName: "bundle") pod "f08e6b54-9291-41eb-9b5c-4265e8e25eb9" (UID: "f08e6b54-9291-41eb-9b5c-4265e8e25eb9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.657915 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-kube-api-access-fb4lj" (OuterVolumeSpecName: "kube-api-access-fb4lj") pod "f08e6b54-9291-41eb-9b5c-4265e8e25eb9" (UID: "f08e6b54-9291-41eb-9b5c-4265e8e25eb9"). InnerVolumeSpecName "kube-api-access-fb4lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.750060 4718 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:20:59 crc kubenswrapper[4718]: I1206 04:20:59.750110 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb4lj\" (UniqueName: \"kubernetes.io/projected/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-kube-api-access-fb4lj\") on node \"crc\" DevicePath \"\"" Dec 06 04:21:00 crc kubenswrapper[4718]: I1206 04:21:00.072142 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-util" (OuterVolumeSpecName: "util") pod "f08e6b54-9291-41eb-9b5c-4265e8e25eb9" (UID: "f08e6b54-9291-41eb-9b5c-4265e8e25eb9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:21:00 crc kubenswrapper[4718]: I1206 04:21:00.154411 4718 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f08e6b54-9291-41eb-9b5c-4265e8e25eb9-util\") on node \"crc\" DevicePath \"\"" Dec 06 04:21:00 crc kubenswrapper[4718]: I1206 04:21:00.259076 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" event={"ID":"f08e6b54-9291-41eb-9b5c-4265e8e25eb9","Type":"ContainerDied","Data":"38f6eec76c1b7b20e2828c671196847a669acaa15210467c3be11ee7c17a963f"} Dec 06 04:21:00 crc kubenswrapper[4718]: I1206 04:21:00.259459 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38f6eec76c1b7b20e2828c671196847a669acaa15210467c3be11ee7c17a963f" Dec 06 04:21:00 crc kubenswrapper[4718]: I1206 04:21:00.259175 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7" Dec 06 04:21:01 crc kubenswrapper[4718]: I1206 04:21:01.268481 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4z8ml" event={"ID":"155c38ca-0913-47f0-9cb1-21063447f752","Type":"ContainerStarted","Data":"9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30"} Dec 06 04:21:01 crc kubenswrapper[4718]: I1206 04:21:01.291022 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4z8ml" podStartSLOduration=2.373788331 podStartE2EDuration="5.290998921s" podCreationTimestamp="2025-12-06 04:20:56 +0000 UTC" firstStartedPulling="2025-12-06 04:20:57.216593737 +0000 UTC m=+846.222298898" lastFinishedPulling="2025-12-06 04:21:00.133804317 +0000 UTC m=+849.139509488" observedRunningTime="2025-12-06 04:21:01.287325755 +0000 UTC m=+850.293030936" watchObservedRunningTime="2025-12-06 04:21:01.290998921 +0000 UTC m=+850.296704122" Dec 06 04:21:06 crc kubenswrapper[4718]: I1206 04:21:06.802901 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:21:06 crc kubenswrapper[4718]: I1206 04:21:06.803535 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:21:06 crc kubenswrapper[4718]: I1206 04:21:06.870484 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:21:07 crc kubenswrapper[4718]: I1206 04:21:07.365765 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.250912 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4z8ml"] Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.251272 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4z8ml" podUID="155c38ca-0913-47f0-9cb1-21063447f752" containerName="registry-server" containerID="cri-o://9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30" gracePeriod=2 Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.499059 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5"] Dec 06 04:21:10 crc kubenswrapper[4718]: E1206 04:21:10.499515 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerName="pull" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.499530 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerName="pull" Dec 06 04:21:10 crc kubenswrapper[4718]: E1206 04:21:10.499547 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerName="extract" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.499553 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerName="extract" Dec 06 04:21:10 crc kubenswrapper[4718]: E1206 04:21:10.499565 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerName="util" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.499570 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerName="util" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.499657 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f08e6b54-9291-41eb-9b5c-4265e8e25eb9" containerName="extract" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.500027 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.502016 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.502040 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-5tc7g" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.502094 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.502162 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.502354 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.517647 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5"] Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.597395 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmnc2\" (UniqueName: \"kubernetes.io/projected/939e915d-3ad4-4a15-bb70-01106cd5cb96-kube-api-access-cmnc2\") pod \"metallb-operator-controller-manager-57d558fdb8-rqqm5\" (UID: \"939e915d-3ad4-4a15-bb70-01106cd5cb96\") " pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.597451 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/939e915d-3ad4-4a15-bb70-01106cd5cb96-webhook-cert\") pod \"metallb-operator-controller-manager-57d558fdb8-rqqm5\" (UID: \"939e915d-3ad4-4a15-bb70-01106cd5cb96\") " pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.597546 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/939e915d-3ad4-4a15-bb70-01106cd5cb96-apiservice-cert\") pod \"metallb-operator-controller-manager-57d558fdb8-rqqm5\" (UID: \"939e915d-3ad4-4a15-bb70-01106cd5cb96\") " pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.698908 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/939e915d-3ad4-4a15-bb70-01106cd5cb96-webhook-cert\") pod \"metallb-operator-controller-manager-57d558fdb8-rqqm5\" (UID: \"939e915d-3ad4-4a15-bb70-01106cd5cb96\") " pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.699004 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/939e915d-3ad4-4a15-bb70-01106cd5cb96-apiservice-cert\") pod \"metallb-operator-controller-manager-57d558fdb8-rqqm5\" (UID: \"939e915d-3ad4-4a15-bb70-01106cd5cb96\") " pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.699034 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmnc2\" (UniqueName: \"kubernetes.io/projected/939e915d-3ad4-4a15-bb70-01106cd5cb96-kube-api-access-cmnc2\") pod \"metallb-operator-controller-manager-57d558fdb8-rqqm5\" (UID: \"939e915d-3ad4-4a15-bb70-01106cd5cb96\") " pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.710993 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/939e915d-3ad4-4a15-bb70-01106cd5cb96-webhook-cert\") pod \"metallb-operator-controller-manager-57d558fdb8-rqqm5\" (UID: \"939e915d-3ad4-4a15-bb70-01106cd5cb96\") " pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.714882 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/939e915d-3ad4-4a15-bb70-01106cd5cb96-apiservice-cert\") pod \"metallb-operator-controller-manager-57d558fdb8-rqqm5\" (UID: \"939e915d-3ad4-4a15-bb70-01106cd5cb96\") " pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.726582 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmnc2\" (UniqueName: \"kubernetes.io/projected/939e915d-3ad4-4a15-bb70-01106cd5cb96-kube-api-access-cmnc2\") pod \"metallb-operator-controller-manager-57d558fdb8-rqqm5\" (UID: \"939e915d-3ad4-4a15-bb70-01106cd5cb96\") " pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:10 crc kubenswrapper[4718]: I1206 04:21:10.813331 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:11 crc kubenswrapper[4718]: I1206 04:21:11.056203 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5"] Dec 06 04:21:11 crc kubenswrapper[4718]: W1206 04:21:11.062564 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod939e915d_3ad4_4a15_bb70_01106cd5cb96.slice/crio-c05ac9d1c1fabf06aaa63ee1a9c6ccc95d51d0b8793a646a8e95f03c9e9dbdbc WatchSource:0}: Error finding container c05ac9d1c1fabf06aaa63ee1a9c6ccc95d51d0b8793a646a8e95f03c9e9dbdbc: Status 404 returned error can't find the container with id c05ac9d1c1fabf06aaa63ee1a9c6ccc95d51d0b8793a646a8e95f03c9e9dbdbc Dec 06 04:21:11 crc kubenswrapper[4718]: I1206 04:21:11.885380 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" event={"ID":"939e915d-3ad4-4a15-bb70-01106cd5cb96","Type":"ContainerStarted","Data":"c05ac9d1c1fabf06aaa63ee1a9c6ccc95d51d0b8793a646a8e95f03c9e9dbdbc"} Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.102128 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt"] Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.103303 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.108763 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-mgrj6" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.108979 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.109054 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.117691 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt"] Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.225244 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5e44d0d-be18-4944-ae68-8758042af15e-webhook-cert\") pod \"metallb-operator-webhook-server-5c49888cd9-w2wwt\" (UID: \"e5e44d0d-be18-4944-ae68-8758042af15e\") " pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.225311 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5e44d0d-be18-4944-ae68-8758042af15e-apiservice-cert\") pod \"metallb-operator-webhook-server-5c49888cd9-w2wwt\" (UID: \"e5e44d0d-be18-4944-ae68-8758042af15e\") " pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.225336 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb7wf\" (UniqueName: \"kubernetes.io/projected/e5e44d0d-be18-4944-ae68-8758042af15e-kube-api-access-bb7wf\") pod \"metallb-operator-webhook-server-5c49888cd9-w2wwt\" (UID: \"e5e44d0d-be18-4944-ae68-8758042af15e\") " pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.326980 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5e44d0d-be18-4944-ae68-8758042af15e-webhook-cert\") pod \"metallb-operator-webhook-server-5c49888cd9-w2wwt\" (UID: \"e5e44d0d-be18-4944-ae68-8758042af15e\") " pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.327065 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5e44d0d-be18-4944-ae68-8758042af15e-apiservice-cert\") pod \"metallb-operator-webhook-server-5c49888cd9-w2wwt\" (UID: \"e5e44d0d-be18-4944-ae68-8758042af15e\") " pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.327095 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb7wf\" (UniqueName: \"kubernetes.io/projected/e5e44d0d-be18-4944-ae68-8758042af15e-kube-api-access-bb7wf\") pod \"metallb-operator-webhook-server-5c49888cd9-w2wwt\" (UID: \"e5e44d0d-be18-4944-ae68-8758042af15e\") " pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.336292 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5e44d0d-be18-4944-ae68-8758042af15e-webhook-cert\") pod \"metallb-operator-webhook-server-5c49888cd9-w2wwt\" (UID: \"e5e44d0d-be18-4944-ae68-8758042af15e\") " pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.336334 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5e44d0d-be18-4944-ae68-8758042af15e-apiservice-cert\") pod \"metallb-operator-webhook-server-5c49888cd9-w2wwt\" (UID: \"e5e44d0d-be18-4944-ae68-8758042af15e\") " pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.343915 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb7wf\" (UniqueName: \"kubernetes.io/projected/e5e44d0d-be18-4944-ae68-8758042af15e-kube-api-access-bb7wf\") pod \"metallb-operator-webhook-server-5c49888cd9-w2wwt\" (UID: \"e5e44d0d-be18-4944-ae68-8758042af15e\") " pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.426140 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.498931 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.630224 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-utilities\") pod \"155c38ca-0913-47f0-9cb1-21063447f752\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.630290 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjvrb\" (UniqueName: \"kubernetes.io/projected/155c38ca-0913-47f0-9cb1-21063447f752-kube-api-access-xjvrb\") pod \"155c38ca-0913-47f0-9cb1-21063447f752\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.630306 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-catalog-content\") pod \"155c38ca-0913-47f0-9cb1-21063447f752\" (UID: \"155c38ca-0913-47f0-9cb1-21063447f752\") " Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.632481 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-utilities" (OuterVolumeSpecName: "utilities") pod "155c38ca-0913-47f0-9cb1-21063447f752" (UID: "155c38ca-0913-47f0-9cb1-21063447f752"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.636433 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/155c38ca-0913-47f0-9cb1-21063447f752-kube-api-access-xjvrb" (OuterVolumeSpecName: "kube-api-access-xjvrb") pod "155c38ca-0913-47f0-9cb1-21063447f752" (UID: "155c38ca-0913-47f0-9cb1-21063447f752"). InnerVolumeSpecName "kube-api-access-xjvrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.731152 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.731731 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjvrb\" (UniqueName: \"kubernetes.io/projected/155c38ca-0913-47f0-9cb1-21063447f752-kube-api-access-xjvrb\") on node \"crc\" DevicePath \"\"" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.751340 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "155c38ca-0913-47f0-9cb1-21063447f752" (UID: "155c38ca-0913-47f0-9cb1-21063447f752"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.832725 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/155c38ca-0913-47f0-9cb1-21063447f752-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.868827 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt"] Dec 06 04:21:12 crc kubenswrapper[4718]: W1206 04:21:12.875271 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5e44d0d_be18_4944_ae68_8758042af15e.slice/crio-c5c36f225f4f5272032d834482aef707c2096732fa5c63c58c6b9823badecd0e WatchSource:0}: Error finding container c5c36f225f4f5272032d834482aef707c2096732fa5c63c58c6b9823badecd0e: Status 404 returned error can't find the container with id c5c36f225f4f5272032d834482aef707c2096732fa5c63c58c6b9823badecd0e Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.895013 4718 generic.go:334] "Generic (PLEG): container finished" podID="155c38ca-0913-47f0-9cb1-21063447f752" containerID="9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30" exitCode=0 Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.895078 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4z8ml" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.895088 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4z8ml" event={"ID":"155c38ca-0913-47f0-9cb1-21063447f752","Type":"ContainerDied","Data":"9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30"} Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.895116 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4z8ml" event={"ID":"155c38ca-0913-47f0-9cb1-21063447f752","Type":"ContainerDied","Data":"b7de719cb3549dbcc90d9cd347964e6eaa90f18fdeb9df3121fe339483faaaf2"} Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.895135 4718 scope.go:117] "RemoveContainer" containerID="9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.897927 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" event={"ID":"e5e44d0d-be18-4944-ae68-8758042af15e","Type":"ContainerStarted","Data":"c5c36f225f4f5272032d834482aef707c2096732fa5c63c58c6b9823badecd0e"} Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.913240 4718 scope.go:117] "RemoveContainer" containerID="fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.935475 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4z8ml"] Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.937432 4718 scope.go:117] "RemoveContainer" containerID="555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.940140 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4z8ml"] Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.956499 4718 scope.go:117] "RemoveContainer" containerID="9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30" Dec 06 04:21:12 crc kubenswrapper[4718]: E1206 04:21:12.957028 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30\": container with ID starting with 9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30 not found: ID does not exist" containerID="9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.957115 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30"} err="failed to get container status \"9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30\": rpc error: code = NotFound desc = could not find container \"9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30\": container with ID starting with 9dc997bb035db152a2083b226963930cacd0137d7ff1df6f2109610aa23bfe30 not found: ID does not exist" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.957166 4718 scope.go:117] "RemoveContainer" containerID="fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f" Dec 06 04:21:12 crc kubenswrapper[4718]: E1206 04:21:12.957667 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f\": container with ID starting with fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f not found: ID does not exist" containerID="fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.957753 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f"} err="failed to get container status \"fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f\": rpc error: code = NotFound desc = could not find container \"fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f\": container with ID starting with fe8aec8dd07fcd537d6aca255efcaecc05c31bf862acd7f868b63039a224cf9f not found: ID does not exist" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.957781 4718 scope.go:117] "RemoveContainer" containerID="555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6" Dec 06 04:21:12 crc kubenswrapper[4718]: E1206 04:21:12.958184 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6\": container with ID starting with 555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6 not found: ID does not exist" containerID="555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6" Dec 06 04:21:12 crc kubenswrapper[4718]: I1206 04:21:12.958267 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6"} err="failed to get container status \"555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6\": rpc error: code = NotFound desc = could not find container \"555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6\": container with ID starting with 555074860c0dd92ae75cab31833ff612f2700a76705df7ebbc9ab8c1cdbd4ad6 not found: ID does not exist" Dec 06 04:21:13 crc kubenswrapper[4718]: I1206 04:21:13.335473 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="155c38ca-0913-47f0-9cb1-21063447f752" path="/var/lib/kubelet/pods/155c38ca-0913-47f0-9cb1-21063447f752/volumes" Dec 06 04:21:14 crc kubenswrapper[4718]: I1206 04:21:14.921196 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" event={"ID":"939e915d-3ad4-4a15-bb70-01106cd5cb96","Type":"ContainerStarted","Data":"828fdc7bf7c9b2a9f27270abbd31c3a78a5da978bd5981ed6432ed0f01d083c5"} Dec 06 04:21:14 crc kubenswrapper[4718]: I1206 04:21:14.921637 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:14 crc kubenswrapper[4718]: I1206 04:21:14.948426 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" podStartSLOduration=2.221241323 podStartE2EDuration="4.948412654s" podCreationTimestamp="2025-12-06 04:21:10 +0000 UTC" firstStartedPulling="2025-12-06 04:21:11.064579944 +0000 UTC m=+860.070285105" lastFinishedPulling="2025-12-06 04:21:13.791751275 +0000 UTC m=+862.797456436" observedRunningTime="2025-12-06 04:21:14.94688579 +0000 UTC m=+863.952590951" watchObservedRunningTime="2025-12-06 04:21:14.948412654 +0000 UTC m=+863.954117815" Dec 06 04:21:17 crc kubenswrapper[4718]: I1206 04:21:17.941595 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" event={"ID":"e5e44d0d-be18-4944-ae68-8758042af15e","Type":"ContainerStarted","Data":"767610bf17459b36b578a93b3add086e6c6311285bd615c34a46d326eac152fe"} Dec 06 04:21:17 crc kubenswrapper[4718]: I1206 04:21:17.942420 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:17 crc kubenswrapper[4718]: I1206 04:21:17.971153 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" podStartSLOduration=1.824205066 podStartE2EDuration="5.971126231s" podCreationTimestamp="2025-12-06 04:21:12 +0000 UTC" firstStartedPulling="2025-12-06 04:21:12.878966883 +0000 UTC m=+861.884672054" lastFinishedPulling="2025-12-06 04:21:17.025888048 +0000 UTC m=+866.031593219" observedRunningTime="2025-12-06 04:21:17.968602168 +0000 UTC m=+866.974307419" watchObservedRunningTime="2025-12-06 04:21:17.971126231 +0000 UTC m=+866.976831432" Dec 06 04:21:32 crc kubenswrapper[4718]: I1206 04:21:32.436002 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5c49888cd9-w2wwt" Dec 06 04:21:50 crc kubenswrapper[4718]: I1206 04:21:50.815288 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-57d558fdb8-rqqm5" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.508810 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-gp6x6"] Dec 06 04:21:51 crc kubenswrapper[4718]: E1206 04:21:51.509526 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155c38ca-0913-47f0-9cb1-21063447f752" containerName="registry-server" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.509567 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="155c38ca-0913-47f0-9cb1-21063447f752" containerName="registry-server" Dec 06 04:21:51 crc kubenswrapper[4718]: E1206 04:21:51.509582 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155c38ca-0913-47f0-9cb1-21063447f752" containerName="extract-content" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.509589 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="155c38ca-0913-47f0-9cb1-21063447f752" containerName="extract-content" Dec 06 04:21:51 crc kubenswrapper[4718]: E1206 04:21:51.509601 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155c38ca-0913-47f0-9cb1-21063447f752" containerName="extract-utilities" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.509609 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="155c38ca-0913-47f0-9cb1-21063447f752" containerName="extract-utilities" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.509748 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="155c38ca-0913-47f0-9cb1-21063447f752" containerName="registry-server" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.511910 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.514358 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-dgphq" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.514582 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.514869 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.517780 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws"] Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.519435 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.524573 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.532200 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws"] Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.582077 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b00eded-0715-4a41-89e9-3f758847a16a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7swws\" (UID: \"0b00eded-0715-4a41-89e9-3f758847a16a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.582124 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e878e223-5476-4e98-ae10-2db9708f0336-metrics-certs\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.582139 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-reloader\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.582160 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-metrics\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.582177 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-frr-sockets\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.582204 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmb9z\" (UniqueName: \"kubernetes.io/projected/e878e223-5476-4e98-ae10-2db9708f0336-kube-api-access-dmb9z\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.582222 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-frr-conf\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.582254 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwj5r\" (UniqueName: \"kubernetes.io/projected/0b00eded-0715-4a41-89e9-3f758847a16a-kube-api-access-jwj5r\") pod \"frr-k8s-webhook-server-7fcb986d4-7swws\" (UID: \"0b00eded-0715-4a41-89e9-3f758847a16a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.582274 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e878e223-5476-4e98-ae10-2db9708f0336-frr-startup\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.620941 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-9wqkt"] Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.622220 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.629487 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.629666 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-r6chz"] Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.629642 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-wxwlc" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.630679 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.630872 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.631099 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.632286 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.643876 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-r6chz"] Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.682812 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-reloader\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.682852 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw4wx\" (UniqueName: \"kubernetes.io/projected/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-kube-api-access-pw4wx\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.682874 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-metrics-certs\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.682893 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-metrics\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.682910 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-frr-sockets\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.683348 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-reloader\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.683311 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-metrics\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.683424 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-frr-sockets\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.683470 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b00ca948-0cd8-43bc-8b38-43f445044e51-cert\") pod \"controller-f8648f98b-r6chz\" (UID: \"b00ca948-0cd8-43bc-8b38-43f445044e51\") " pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.683498 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmb9z\" (UniqueName: \"kubernetes.io/projected/e878e223-5476-4e98-ae10-2db9708f0336-kube-api-access-dmb9z\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.683520 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-frr-conf\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.683538 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwj5r\" (UniqueName: \"kubernetes.io/projected/0b00eded-0715-4a41-89e9-3f758847a16a-kube-api-access-jwj5r\") pod \"frr-k8s-webhook-server-7fcb986d4-7swws\" (UID: \"0b00eded-0715-4a41-89e9-3f758847a16a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.683994 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e878e223-5476-4e98-ae10-2db9708f0336-frr-startup\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.683830 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e878e223-5476-4e98-ae10-2db9708f0336-frr-conf\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.684735 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e878e223-5476-4e98-ae10-2db9708f0336-frr-startup\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.684023 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-metallb-excludel2\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.684827 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b00ca948-0cd8-43bc-8b38-43f445044e51-metrics-certs\") pod \"controller-f8648f98b-r6chz\" (UID: \"b00ca948-0cd8-43bc-8b38-43f445044e51\") " pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.684842 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-memberlist\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.684904 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x78mk\" (UniqueName: \"kubernetes.io/projected/b00ca948-0cd8-43bc-8b38-43f445044e51-kube-api-access-x78mk\") pod \"controller-f8648f98b-r6chz\" (UID: \"b00ca948-0cd8-43bc-8b38-43f445044e51\") " pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.684989 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b00eded-0715-4a41-89e9-3f758847a16a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7swws\" (UID: \"0b00eded-0715-4a41-89e9-3f758847a16a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.685006 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e878e223-5476-4e98-ae10-2db9708f0336-metrics-certs\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: E1206 04:21:51.685550 4718 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 06 04:21:51 crc kubenswrapper[4718]: E1206 04:21:51.685608 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b00eded-0715-4a41-89e9-3f758847a16a-cert podName:0b00eded-0715-4a41-89e9-3f758847a16a nodeName:}" failed. No retries permitted until 2025-12-06 04:21:52.185580015 +0000 UTC m=+901.191285176 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0b00eded-0715-4a41-89e9-3f758847a16a-cert") pod "frr-k8s-webhook-server-7fcb986d4-7swws" (UID: "0b00eded-0715-4a41-89e9-3f758847a16a") : secret "frr-k8s-webhook-server-cert" not found Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.696357 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e878e223-5476-4e98-ae10-2db9708f0336-metrics-certs\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.700787 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmb9z\" (UniqueName: \"kubernetes.io/projected/e878e223-5476-4e98-ae10-2db9708f0336-kube-api-access-dmb9z\") pod \"frr-k8s-gp6x6\" (UID: \"e878e223-5476-4e98-ae10-2db9708f0336\") " pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.707394 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwj5r\" (UniqueName: \"kubernetes.io/projected/0b00eded-0715-4a41-89e9-3f758847a16a-kube-api-access-jwj5r\") pod \"frr-k8s-webhook-server-7fcb986d4-7swws\" (UID: \"0b00eded-0715-4a41-89e9-3f758847a16a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.787374 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b00ca948-0cd8-43bc-8b38-43f445044e51-metrics-certs\") pod \"controller-f8648f98b-r6chz\" (UID: \"b00ca948-0cd8-43bc-8b38-43f445044e51\") " pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.787417 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-memberlist\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.787450 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x78mk\" (UniqueName: \"kubernetes.io/projected/b00ca948-0cd8-43bc-8b38-43f445044e51-kube-api-access-x78mk\") pod \"controller-f8648f98b-r6chz\" (UID: \"b00ca948-0cd8-43bc-8b38-43f445044e51\") " pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.787485 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw4wx\" (UniqueName: \"kubernetes.io/projected/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-kube-api-access-pw4wx\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.787502 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-metrics-certs\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.787526 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b00ca948-0cd8-43bc-8b38-43f445044e51-cert\") pod \"controller-f8648f98b-r6chz\" (UID: \"b00ca948-0cd8-43bc-8b38-43f445044e51\") " pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.787558 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-metallb-excludel2\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.788215 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-metallb-excludel2\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: E1206 04:21:51.788528 4718 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 06 04:21:51 crc kubenswrapper[4718]: E1206 04:21:51.788572 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-metrics-certs podName:1bc1c47e-1be5-4a69-96cf-fa5c59004f8b nodeName:}" failed. No retries permitted until 2025-12-06 04:21:52.288560328 +0000 UTC m=+901.294265489 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-metrics-certs") pod "speaker-9wqkt" (UID: "1bc1c47e-1be5-4a69-96cf-fa5c59004f8b") : secret "speaker-certs-secret" not found Dec 06 04:21:51 crc kubenswrapper[4718]: E1206 04:21:51.788695 4718 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 04:21:51 crc kubenswrapper[4718]: E1206 04:21:51.788722 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-memberlist podName:1bc1c47e-1be5-4a69-96cf-fa5c59004f8b nodeName:}" failed. No retries permitted until 2025-12-06 04:21:52.288715543 +0000 UTC m=+901.294420704 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-memberlist") pod "speaker-9wqkt" (UID: "1bc1c47e-1be5-4a69-96cf-fa5c59004f8b") : secret "metallb-memberlist" not found Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.792721 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b00ca948-0cd8-43bc-8b38-43f445044e51-metrics-certs\") pod \"controller-f8648f98b-r6chz\" (UID: \"b00ca948-0cd8-43bc-8b38-43f445044e51\") " pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.795309 4718 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.801895 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b00ca948-0cd8-43bc-8b38-43f445044e51-cert\") pod \"controller-f8648f98b-r6chz\" (UID: \"b00ca948-0cd8-43bc-8b38-43f445044e51\") " pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.815742 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw4wx\" (UniqueName: \"kubernetes.io/projected/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-kube-api-access-pw4wx\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.820121 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x78mk\" (UniqueName: \"kubernetes.io/projected/b00ca948-0cd8-43bc-8b38-43f445044e51-kube-api-access-x78mk\") pod \"controller-f8648f98b-r6chz\" (UID: \"b00ca948-0cd8-43bc-8b38-43f445044e51\") " pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.840528 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:21:51 crc kubenswrapper[4718]: I1206 04:21:51.958255 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:52 crc kubenswrapper[4718]: I1206 04:21:52.152802 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-r6chz"] Dec 06 04:21:52 crc kubenswrapper[4718]: W1206 04:21:52.165407 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb00ca948_0cd8_43bc_8b38_43f445044e51.slice/crio-ab756f94593aa50b96ff202d2634d18e628bbf315c0c4ef8b17f45d404c4eb3a WatchSource:0}: Error finding container ab756f94593aa50b96ff202d2634d18e628bbf315c0c4ef8b17f45d404c4eb3a: Status 404 returned error can't find the container with id ab756f94593aa50b96ff202d2634d18e628bbf315c0c4ef8b17f45d404c4eb3a Dec 06 04:21:52 crc kubenswrapper[4718]: I1206 04:21:52.193354 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b00eded-0715-4a41-89e9-3f758847a16a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7swws\" (UID: \"0b00eded-0715-4a41-89e9-3f758847a16a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:21:52 crc kubenswrapper[4718]: I1206 04:21:52.199858 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b00eded-0715-4a41-89e9-3f758847a16a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-7swws\" (UID: \"0b00eded-0715-4a41-89e9-3f758847a16a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:21:52 crc kubenswrapper[4718]: I1206 04:21:52.294595 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-metrics-certs\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:52 crc kubenswrapper[4718]: I1206 04:21:52.294692 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-memberlist\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:52 crc kubenswrapper[4718]: E1206 04:21:52.294812 4718 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 04:21:52 crc kubenswrapper[4718]: E1206 04:21:52.294869 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-memberlist podName:1bc1c47e-1be5-4a69-96cf-fa5c59004f8b nodeName:}" failed. No retries permitted until 2025-12-06 04:21:53.294855971 +0000 UTC m=+902.300561132 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-memberlist") pod "speaker-9wqkt" (UID: "1bc1c47e-1be5-4a69-96cf-fa5c59004f8b") : secret "metallb-memberlist" not found Dec 06 04:21:52 crc kubenswrapper[4718]: I1206 04:21:52.300791 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-metrics-certs\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:52 crc kubenswrapper[4718]: I1206 04:21:52.456883 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:21:52 crc kubenswrapper[4718]: I1206 04:21:52.922077 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws"] Dec 06 04:21:52 crc kubenswrapper[4718]: W1206 04:21:52.925432 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b00eded_0715_4a41_89e9_3f758847a16a.slice/crio-caf4f01558f7ec691a2be8e86d9ae3c1d989ee0b49385f265caf01d20b4fd3b1 WatchSource:0}: Error finding container caf4f01558f7ec691a2be8e86d9ae3c1d989ee0b49385f265caf01d20b4fd3b1: Status 404 returned error can't find the container with id caf4f01558f7ec691a2be8e86d9ae3c1d989ee0b49385f265caf01d20b4fd3b1 Dec 06 04:21:53 crc kubenswrapper[4718]: I1206 04:21:53.172295 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerStarted","Data":"31d52b5e25ab335c47a2e6e2e30237c7fa1276a0e52be2c6a304d594b877e3bd"} Dec 06 04:21:53 crc kubenswrapper[4718]: I1206 04:21:53.174046 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-r6chz" event={"ID":"b00ca948-0cd8-43bc-8b38-43f445044e51","Type":"ContainerStarted","Data":"8b24c3fa0a4fd901583b6e08a273543cba9874bc8d0ecd19d1056eea8e562fc1"} Dec 06 04:21:53 crc kubenswrapper[4718]: I1206 04:21:53.174074 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-r6chz" event={"ID":"b00ca948-0cd8-43bc-8b38-43f445044e51","Type":"ContainerStarted","Data":"ab756f94593aa50b96ff202d2634d18e628bbf315c0c4ef8b17f45d404c4eb3a"} Dec 06 04:21:53 crc kubenswrapper[4718]: I1206 04:21:53.175353 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" event={"ID":"0b00eded-0715-4a41-89e9-3f758847a16a","Type":"ContainerStarted","Data":"caf4f01558f7ec691a2be8e86d9ae3c1d989ee0b49385f265caf01d20b4fd3b1"} Dec 06 04:21:53 crc kubenswrapper[4718]: I1206 04:21:53.308871 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-memberlist\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:53 crc kubenswrapper[4718]: I1206 04:21:53.316709 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bc1c47e-1be5-4a69-96cf-fa5c59004f8b-memberlist\") pod \"speaker-9wqkt\" (UID: \"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b\") " pod="metallb-system/speaker-9wqkt" Dec 06 04:21:53 crc kubenswrapper[4718]: I1206 04:21:53.445755 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9wqkt" Dec 06 04:21:53 crc kubenswrapper[4718]: W1206 04:21:53.465019 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bc1c47e_1be5_4a69_96cf_fa5c59004f8b.slice/crio-0da39a58e0f9e35be00e1dad22033900b39b0c0938b49bd53e834d01a503de42 WatchSource:0}: Error finding container 0da39a58e0f9e35be00e1dad22033900b39b0c0938b49bd53e834d01a503de42: Status 404 returned error can't find the container with id 0da39a58e0f9e35be00e1dad22033900b39b0c0938b49bd53e834d01a503de42 Dec 06 04:21:54 crc kubenswrapper[4718]: I1206 04:21:54.183608 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9wqkt" event={"ID":"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b","Type":"ContainerStarted","Data":"c587afa6e7d54404c8662663a8c6490f76aed1bb359710909933545152dd08c7"} Dec 06 04:21:54 crc kubenswrapper[4718]: I1206 04:21:54.183835 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9wqkt" event={"ID":"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b","Type":"ContainerStarted","Data":"0da39a58e0f9e35be00e1dad22033900b39b0c0938b49bd53e834d01a503de42"} Dec 06 04:21:56 crc kubenswrapper[4718]: I1206 04:21:56.198723 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-r6chz" event={"ID":"b00ca948-0cd8-43bc-8b38-43f445044e51","Type":"ContainerStarted","Data":"c987537f8f16d708133c8a2044673b1f643725e4422f642bcb6bf06150a2c6a8"} Dec 06 04:21:56 crc kubenswrapper[4718]: I1206 04:21:56.199041 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:21:56 crc kubenswrapper[4718]: I1206 04:21:56.213692 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-r6chz" podStartSLOduration=2.060990512 podStartE2EDuration="5.213678026s" podCreationTimestamp="2025-12-06 04:21:51 +0000 UTC" firstStartedPulling="2025-12-06 04:21:52.855750874 +0000 UTC m=+901.861456045" lastFinishedPulling="2025-12-06 04:21:56.008438398 +0000 UTC m=+905.014143559" observedRunningTime="2025-12-06 04:21:56.213001226 +0000 UTC m=+905.218706387" watchObservedRunningTime="2025-12-06 04:21:56.213678026 +0000 UTC m=+905.219383187" Dec 06 04:21:57 crc kubenswrapper[4718]: I1206 04:21:57.211355 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9wqkt" event={"ID":"1bc1c47e-1be5-4a69-96cf-fa5c59004f8b","Type":"ContainerStarted","Data":"62a54971253ffba6bc27c42097337bfb505184dffd39bd8ce86b450a55bb3224"} Dec 06 04:21:57 crc kubenswrapper[4718]: I1206 04:21:57.211833 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-9wqkt" Dec 06 04:21:57 crc kubenswrapper[4718]: I1206 04:21:57.250640 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-9wqkt" podStartSLOduration=3.9222449 podStartE2EDuration="6.250620865s" podCreationTimestamp="2025-12-06 04:21:51 +0000 UTC" firstStartedPulling="2025-12-06 04:21:53.687404595 +0000 UTC m=+902.693109756" lastFinishedPulling="2025-12-06 04:21:56.01578056 +0000 UTC m=+905.021485721" observedRunningTime="2025-12-06 04:21:57.24731745 +0000 UTC m=+906.253022611" watchObservedRunningTime="2025-12-06 04:21:57.250620865 +0000 UTC m=+906.256326046" Dec 06 04:22:00 crc kubenswrapper[4718]: I1206 04:22:00.228153 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" event={"ID":"0b00eded-0715-4a41-89e9-3f758847a16a","Type":"ContainerStarted","Data":"43ccd108f2a624e4a3084697c050656cceba18043f0c691959ddef8d4f6644a2"} Dec 06 04:22:00 crc kubenswrapper[4718]: I1206 04:22:00.229503 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:22:00 crc kubenswrapper[4718]: I1206 04:22:00.230052 4718 generic.go:334] "Generic (PLEG): container finished" podID="e878e223-5476-4e98-ae10-2db9708f0336" containerID="e7661b836233640914a400de030c9704a26938a93f389d8f4cb2060c76b688bb" exitCode=0 Dec 06 04:22:00 crc kubenswrapper[4718]: I1206 04:22:00.230091 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerDied","Data":"e7661b836233640914a400de030c9704a26938a93f389d8f4cb2060c76b688bb"} Dec 06 04:22:00 crc kubenswrapper[4718]: I1206 04:22:00.249748 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" podStartSLOduration=2.951413547 podStartE2EDuration="9.249732658s" podCreationTimestamp="2025-12-06 04:21:51 +0000 UTC" firstStartedPulling="2025-12-06 04:21:52.92871516 +0000 UTC m=+901.934420331" lastFinishedPulling="2025-12-06 04:21:59.227034281 +0000 UTC m=+908.232739442" observedRunningTime="2025-12-06 04:22:00.249584664 +0000 UTC m=+909.255289835" watchObservedRunningTime="2025-12-06 04:22:00.249732658 +0000 UTC m=+909.255437819" Dec 06 04:22:01 crc kubenswrapper[4718]: I1206 04:22:01.244386 4718 generic.go:334] "Generic (PLEG): container finished" podID="e878e223-5476-4e98-ae10-2db9708f0336" containerID="d25d7d6f7d3b588a6f24855864f68ca500f62a6f86b5da9599cdfa95c052f9a9" exitCode=0 Dec 06 04:22:01 crc kubenswrapper[4718]: I1206 04:22:01.244467 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerDied","Data":"d25d7d6f7d3b588a6f24855864f68ca500f62a6f86b5da9599cdfa95c052f9a9"} Dec 06 04:22:02 crc kubenswrapper[4718]: I1206 04:22:02.253667 4718 generic.go:334] "Generic (PLEG): container finished" podID="e878e223-5476-4e98-ae10-2db9708f0336" containerID="04ad7337f35d5e831ee22574efaf44221e39fc8ddf058361f3d7d055cab06f04" exitCode=0 Dec 06 04:22:02 crc kubenswrapper[4718]: I1206 04:22:02.253807 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerDied","Data":"04ad7337f35d5e831ee22574efaf44221e39fc8ddf058361f3d7d055cab06f04"} Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.269705 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerStarted","Data":"5d0b9c014b5c7120dff1902847381e43ad0da0ac05a45e4ba712af4fe85e7553"} Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.272529 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerStarted","Data":"1ffcec6a14e0e48992a516ac08526e35524c4661f00ca74d3e9e5b8042ddfb4d"} Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.272570 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerStarted","Data":"af4e5a3af2656f137fd3d5f6699d49022d42b9f08677fc9add85b4001fdcc857"} Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.272584 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerStarted","Data":"89954b6e3a5a255e91d05e2e55891c254d1f691664076eeba37bec9f51b67b93"} Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.272596 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerStarted","Data":"b9608fb9946917417b6356031b68f6b000b9a5c81121605b1f7aafaac6208838"} Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.449758 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-9wqkt" Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.829276 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-4x985"] Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.829974 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4x985" Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.838306 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-5khtq" Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.847994 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-4x985"] Dec 06 04:22:03 crc kubenswrapper[4718]: I1206 04:22:03.950174 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dwhk\" (UniqueName: \"kubernetes.io/projected/97dc9f10-c57a-4485-a4d6-b7692c14d00f-kube-api-access-2dwhk\") pod \"infra-operator-index-4x985\" (UID: \"97dc9f10-c57a-4485-a4d6-b7692c14d00f\") " pod="openstack-operators/infra-operator-index-4x985" Dec 06 04:22:04 crc kubenswrapper[4718]: I1206 04:22:04.051859 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dwhk\" (UniqueName: \"kubernetes.io/projected/97dc9f10-c57a-4485-a4d6-b7692c14d00f-kube-api-access-2dwhk\") pod \"infra-operator-index-4x985\" (UID: \"97dc9f10-c57a-4485-a4d6-b7692c14d00f\") " pod="openstack-operators/infra-operator-index-4x985" Dec 06 04:22:04 crc kubenswrapper[4718]: I1206 04:22:04.086675 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dwhk\" (UniqueName: \"kubernetes.io/projected/97dc9f10-c57a-4485-a4d6-b7692c14d00f-kube-api-access-2dwhk\") pod \"infra-operator-index-4x985\" (UID: \"97dc9f10-c57a-4485-a4d6-b7692c14d00f\") " pod="openstack-operators/infra-operator-index-4x985" Dec 06 04:22:04 crc kubenswrapper[4718]: I1206 04:22:04.143983 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4x985" Dec 06 04:22:04 crc kubenswrapper[4718]: I1206 04:22:04.287028 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gp6x6" event={"ID":"e878e223-5476-4e98-ae10-2db9708f0336","Type":"ContainerStarted","Data":"dcf4d9239eabab0e287f9420a13e9c86133904bec95cacead468b86721d8877a"} Dec 06 04:22:04 crc kubenswrapper[4718]: I1206 04:22:04.287306 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:22:04 crc kubenswrapper[4718]: I1206 04:22:04.311418 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-gp6x6" podStartSLOduration=6.901402314 podStartE2EDuration="13.311400522s" podCreationTimestamp="2025-12-06 04:21:51 +0000 UTC" firstStartedPulling="2025-12-06 04:21:52.840032049 +0000 UTC m=+901.845737210" lastFinishedPulling="2025-12-06 04:21:59.250030257 +0000 UTC m=+908.255735418" observedRunningTime="2025-12-06 04:22:04.307593423 +0000 UTC m=+913.313298594" watchObservedRunningTime="2025-12-06 04:22:04.311400522 +0000 UTC m=+913.317105683" Dec 06 04:22:04 crc kubenswrapper[4718]: I1206 04:22:04.348570 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-4x985"] Dec 06 04:22:05 crc kubenswrapper[4718]: I1206 04:22:05.301379 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4x985" event={"ID":"97dc9f10-c57a-4485-a4d6-b7692c14d00f","Type":"ContainerStarted","Data":"916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5"} Dec 06 04:22:05 crc kubenswrapper[4718]: I1206 04:22:05.301634 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4x985" event={"ID":"97dc9f10-c57a-4485-a4d6-b7692c14d00f","Type":"ContainerStarted","Data":"e4e1fcc21624085b084421e7b22cdf0ce98fc0121f9ecacd9b6bc9d874032527"} Dec 06 04:22:05 crc kubenswrapper[4718]: I1206 04:22:05.320530 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-4x985" podStartSLOduration=1.57667402 podStartE2EDuration="2.320514176s" podCreationTimestamp="2025-12-06 04:22:03 +0000 UTC" firstStartedPulling="2025-12-06 04:22:04.382366029 +0000 UTC m=+913.388071200" lastFinishedPulling="2025-12-06 04:22:05.126206155 +0000 UTC m=+914.131911356" observedRunningTime="2025-12-06 04:22:05.318084865 +0000 UTC m=+914.323790026" watchObservedRunningTime="2025-12-06 04:22:05.320514176 +0000 UTC m=+914.326219347" Dec 06 04:22:06 crc kubenswrapper[4718]: I1206 04:22:06.629790 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-4x985"] Dec 06 04:22:06 crc kubenswrapper[4718]: I1206 04:22:06.841746 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:22:06 crc kubenswrapper[4718]: I1206 04:22:06.890403 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:22:07 crc kubenswrapper[4718]: I1206 04:22:07.233407 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-8nhbw"] Dec 06 04:22:07 crc kubenswrapper[4718]: I1206 04:22:07.234191 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:22:07 crc kubenswrapper[4718]: I1206 04:22:07.244109 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-8nhbw"] Dec 06 04:22:07 crc kubenswrapper[4718]: I1206 04:22:07.303968 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkdcd\" (UniqueName: \"kubernetes.io/projected/4ca99b94-f080-4ebc-9010-6aa0a0bcfac4-kube-api-access-nkdcd\") pod \"infra-operator-index-8nhbw\" (UID: \"4ca99b94-f080-4ebc-9010-6aa0a0bcfac4\") " pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:22:07 crc kubenswrapper[4718]: I1206 04:22:07.311384 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-4x985" podUID="97dc9f10-c57a-4485-a4d6-b7692c14d00f" containerName="registry-server" containerID="cri-o://916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5" gracePeriod=2 Dec 06 04:22:07 crc kubenswrapper[4718]: I1206 04:22:07.405148 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkdcd\" (UniqueName: \"kubernetes.io/projected/4ca99b94-f080-4ebc-9010-6aa0a0bcfac4-kube-api-access-nkdcd\") pod \"infra-operator-index-8nhbw\" (UID: \"4ca99b94-f080-4ebc-9010-6aa0a0bcfac4\") " pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:22:07 crc kubenswrapper[4718]: I1206 04:22:07.431070 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkdcd\" (UniqueName: \"kubernetes.io/projected/4ca99b94-f080-4ebc-9010-6aa0a0bcfac4-kube-api-access-nkdcd\") pod \"infra-operator-index-8nhbw\" (UID: \"4ca99b94-f080-4ebc-9010-6aa0a0bcfac4\") " pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:22:08 crc kubenswrapper[4718]: I1206 04:22:08.157074 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:22:08 crc kubenswrapper[4718]: I1206 04:22:08.450262 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-8nhbw"] Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.053779 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4x985" Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.070706 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dwhk\" (UniqueName: \"kubernetes.io/projected/97dc9f10-c57a-4485-a4d6-b7692c14d00f-kube-api-access-2dwhk\") pod \"97dc9f10-c57a-4485-a4d6-b7692c14d00f\" (UID: \"97dc9f10-c57a-4485-a4d6-b7692c14d00f\") " Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.084226 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97dc9f10-c57a-4485-a4d6-b7692c14d00f-kube-api-access-2dwhk" (OuterVolumeSpecName: "kube-api-access-2dwhk") pod "97dc9f10-c57a-4485-a4d6-b7692c14d00f" (UID: "97dc9f10-c57a-4485-a4d6-b7692c14d00f"). InnerVolumeSpecName "kube-api-access-2dwhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.172541 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dwhk\" (UniqueName: \"kubernetes.io/projected/97dc9f10-c57a-4485-a4d6-b7692c14d00f-kube-api-access-2dwhk\") on node \"crc\" DevicePath \"\"" Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.345654 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8nhbw" event={"ID":"4ca99b94-f080-4ebc-9010-6aa0a0bcfac4","Type":"ContainerStarted","Data":"95deed81ab3e6f1a9ee4eca0594e53dbcbbc93fa0b1bdbb10f574c8c5df70273"} Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.345688 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8nhbw" event={"ID":"4ca99b94-f080-4ebc-9010-6aa0a0bcfac4","Type":"ContainerStarted","Data":"4a57a2d515794ce33486d1240391e2e3f9fb0efb79a4e819906eb2ad788a231a"} Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.356323 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4x985" Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.356355 4718 generic.go:334] "Generic (PLEG): container finished" podID="97dc9f10-c57a-4485-a4d6-b7692c14d00f" containerID="916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5" exitCode=0 Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.356390 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4x985" event={"ID":"97dc9f10-c57a-4485-a4d6-b7692c14d00f","Type":"ContainerDied","Data":"916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5"} Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.356453 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4x985" event={"ID":"97dc9f10-c57a-4485-a4d6-b7692c14d00f","Type":"ContainerDied","Data":"e4e1fcc21624085b084421e7b22cdf0ce98fc0121f9ecacd9b6bc9d874032527"} Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.356473 4718 scope.go:117] "RemoveContainer" containerID="916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5" Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.373142 4718 scope.go:117] "RemoveContainer" containerID="916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5" Dec 06 04:22:09 crc kubenswrapper[4718]: E1206 04:22:09.373568 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5\": container with ID starting with 916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5 not found: ID does not exist" containerID="916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5" Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.373625 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5"} err="failed to get container status \"916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5\": rpc error: code = NotFound desc = could not find container \"916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5\": container with ID starting with 916ebf90319cb803eca4c205031d68b095cbd8a5304066f5f898dd834afc2bf5 not found: ID does not exist" Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.379096 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-8nhbw" podStartSLOduration=1.940834711 podStartE2EDuration="2.379085471s" podCreationTimestamp="2025-12-06 04:22:07 +0000 UTC" firstStartedPulling="2025-12-06 04:22:08.461419407 +0000 UTC m=+917.467124568" lastFinishedPulling="2025-12-06 04:22:08.899670167 +0000 UTC m=+917.905375328" observedRunningTime="2025-12-06 04:22:09.378753031 +0000 UTC m=+918.384458192" watchObservedRunningTime="2025-12-06 04:22:09.379085471 +0000 UTC m=+918.384790632" Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.393992 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-4x985"] Dec 06 04:22:09 crc kubenswrapper[4718]: I1206 04:22:09.398000 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-4x985"] Dec 06 04:22:11 crc kubenswrapper[4718]: I1206 04:22:11.341064 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97dc9f10-c57a-4485-a4d6-b7692c14d00f" path="/var/lib/kubelet/pods/97dc9f10-c57a-4485-a4d6-b7692c14d00f/volumes" Dec 06 04:22:11 crc kubenswrapper[4718]: I1206 04:22:11.965276 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-r6chz" Dec 06 04:22:12 crc kubenswrapper[4718]: I1206 04:22:12.467258 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-7swws" Dec 06 04:22:18 crc kubenswrapper[4718]: I1206 04:22:18.158504 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:22:18 crc kubenswrapper[4718]: I1206 04:22:18.159783 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:22:18 crc kubenswrapper[4718]: I1206 04:22:18.191808 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:22:18 crc kubenswrapper[4718]: I1206 04:22:18.460951 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.301058 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9"] Dec 06 04:22:20 crc kubenswrapper[4718]: E1206 04:22:20.301847 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97dc9f10-c57a-4485-a4d6-b7692c14d00f" containerName="registry-server" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.301920 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="97dc9f10-c57a-4485-a4d6-b7692c14d00f" containerName="registry-server" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.302072 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="97dc9f10-c57a-4485-a4d6-b7692c14d00f" containerName="registry-server" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.302841 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.308140 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-69frt" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.316630 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9"] Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.322421 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvvfk\" (UniqueName: \"kubernetes.io/projected/766a146c-c056-4843-be3b-fd5d4572624c-kube-api-access-jvvfk\") pod \"3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.322510 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-bundle\") pod \"3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.322624 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-util\") pod \"3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.423661 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvvfk\" (UniqueName: \"kubernetes.io/projected/766a146c-c056-4843-be3b-fd5d4572624c-kube-api-access-jvvfk\") pod \"3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.423932 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-bundle\") pod \"3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.424014 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-util\") pod \"3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.424480 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-bundle\") pod \"3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.424587 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-util\") pod \"3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.446272 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvvfk\" (UniqueName: \"kubernetes.io/projected/766a146c-c056-4843-be3b-fd5d4572624c-kube-api-access-jvvfk\") pod \"3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.622306 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:20 crc kubenswrapper[4718]: I1206 04:22:20.833062 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9"] Dec 06 04:22:20 crc kubenswrapper[4718]: W1206 04:22:20.835210 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod766a146c_c056_4843_be3b_fd5d4572624c.slice/crio-4e733bbec9e87dcfc65b9665241e690b4de74617829ca1eb77fa66226c0a3aa4 WatchSource:0}: Error finding container 4e733bbec9e87dcfc65b9665241e690b4de74617829ca1eb77fa66226c0a3aa4: Status 404 returned error can't find the container with id 4e733bbec9e87dcfc65b9665241e690b4de74617829ca1eb77fa66226c0a3aa4 Dec 06 04:22:21 crc kubenswrapper[4718]: I1206 04:22:21.436891 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" event={"ID":"766a146c-c056-4843-be3b-fd5d4572624c","Type":"ContainerStarted","Data":"4e733bbec9e87dcfc65b9665241e690b4de74617829ca1eb77fa66226c0a3aa4"} Dec 06 04:22:21 crc kubenswrapper[4718]: I1206 04:22:21.847530 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-gp6x6" Dec 06 04:22:23 crc kubenswrapper[4718]: I1206 04:22:23.454194 4718 generic.go:334] "Generic (PLEG): container finished" podID="766a146c-c056-4843-be3b-fd5d4572624c" containerID="b9bb347c8c4f66de0bbffb988d1988ce2defc3534b84539638e8a0d80a129ceb" exitCode=0 Dec 06 04:22:23 crc kubenswrapper[4718]: I1206 04:22:23.454267 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" event={"ID":"766a146c-c056-4843-be3b-fd5d4572624c","Type":"ContainerDied","Data":"b9bb347c8c4f66de0bbffb988d1988ce2defc3534b84539638e8a0d80a129ceb"} Dec 06 04:22:24 crc kubenswrapper[4718]: I1206 04:22:24.461824 4718 generic.go:334] "Generic (PLEG): container finished" podID="766a146c-c056-4843-be3b-fd5d4572624c" containerID="2cdfdb17b2603828d33a7279cdb768c507d3dadec658af3c6a3cf287a1d4df22" exitCode=0 Dec 06 04:22:24 crc kubenswrapper[4718]: I1206 04:22:24.462143 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" event={"ID":"766a146c-c056-4843-be3b-fd5d4572624c","Type":"ContainerDied","Data":"2cdfdb17b2603828d33a7279cdb768c507d3dadec658af3c6a3cf287a1d4df22"} Dec 06 04:22:25 crc kubenswrapper[4718]: I1206 04:22:25.470916 4718 generic.go:334] "Generic (PLEG): container finished" podID="766a146c-c056-4843-be3b-fd5d4572624c" containerID="21cf09f334cc01a5c689c212a101b15b4d768daa5648ebbe24dc25314d337ab7" exitCode=0 Dec 06 04:22:25 crc kubenswrapper[4718]: I1206 04:22:25.471061 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" event={"ID":"766a146c-c056-4843-be3b-fd5d4572624c","Type":"ContainerDied","Data":"21cf09f334cc01a5c689c212a101b15b4d768daa5648ebbe24dc25314d337ab7"} Dec 06 04:22:26 crc kubenswrapper[4718]: I1206 04:22:26.736913 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:26 crc kubenswrapper[4718]: I1206 04:22:26.933726 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-util\") pod \"766a146c-c056-4843-be3b-fd5d4572624c\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " Dec 06 04:22:26 crc kubenswrapper[4718]: I1206 04:22:26.933795 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvvfk\" (UniqueName: \"kubernetes.io/projected/766a146c-c056-4843-be3b-fd5d4572624c-kube-api-access-jvvfk\") pod \"766a146c-c056-4843-be3b-fd5d4572624c\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " Dec 06 04:22:26 crc kubenswrapper[4718]: I1206 04:22:26.933836 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-bundle\") pod \"766a146c-c056-4843-be3b-fd5d4572624c\" (UID: \"766a146c-c056-4843-be3b-fd5d4572624c\") " Dec 06 04:22:26 crc kubenswrapper[4718]: I1206 04:22:26.935656 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-bundle" (OuterVolumeSpecName: "bundle") pod "766a146c-c056-4843-be3b-fd5d4572624c" (UID: "766a146c-c056-4843-be3b-fd5d4572624c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:22:26 crc kubenswrapper[4718]: I1206 04:22:26.942606 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/766a146c-c056-4843-be3b-fd5d4572624c-kube-api-access-jvvfk" (OuterVolumeSpecName: "kube-api-access-jvvfk") pod "766a146c-c056-4843-be3b-fd5d4572624c" (UID: "766a146c-c056-4843-be3b-fd5d4572624c"). InnerVolumeSpecName "kube-api-access-jvvfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:22:26 crc kubenswrapper[4718]: I1206 04:22:26.955299 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-util" (OuterVolumeSpecName: "util") pod "766a146c-c056-4843-be3b-fd5d4572624c" (UID: "766a146c-c056-4843-be3b-fd5d4572624c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:22:27 crc kubenswrapper[4718]: I1206 04:22:27.035348 4718 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:22:27 crc kubenswrapper[4718]: I1206 04:22:27.035392 4718 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/766a146c-c056-4843-be3b-fd5d4572624c-util\") on node \"crc\" DevicePath \"\"" Dec 06 04:22:27 crc kubenswrapper[4718]: I1206 04:22:27.035410 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvvfk\" (UniqueName: \"kubernetes.io/projected/766a146c-c056-4843-be3b-fd5d4572624c-kube-api-access-jvvfk\") on node \"crc\" DevicePath \"\"" Dec 06 04:22:27 crc kubenswrapper[4718]: I1206 04:22:27.486585 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" event={"ID":"766a146c-c056-4843-be3b-fd5d4572624c","Type":"ContainerDied","Data":"4e733bbec9e87dcfc65b9665241e690b4de74617829ca1eb77fa66226c0a3aa4"} Dec 06 04:22:27 crc kubenswrapper[4718]: I1206 04:22:27.486647 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e733bbec9e87dcfc65b9665241e690b4de74617829ca1eb77fa66226c0a3aa4" Dec 06 04:22:27 crc kubenswrapper[4718]: I1206 04:22:27.486741 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.641403 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Dec 06 04:22:36 crc kubenswrapper[4718]: E1206 04:22:36.642598 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="766a146c-c056-4843-be3b-fd5d4572624c" containerName="extract" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.642631 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="766a146c-c056-4843-be3b-fd5d4572624c" containerName="extract" Dec 06 04:22:36 crc kubenswrapper[4718]: E1206 04:22:36.642656 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="766a146c-c056-4843-be3b-fd5d4572624c" containerName="util" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.642672 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="766a146c-c056-4843-be3b-fd5d4572624c" containerName="util" Dec 06 04:22:36 crc kubenswrapper[4718]: E1206 04:22:36.642728 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="766a146c-c056-4843-be3b-fd5d4572624c" containerName="pull" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.642744 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="766a146c-c056-4843-be3b-fd5d4572624c" containerName="pull" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.643008 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="766a146c-c056-4843-be3b-fd5d4572624c" containerName="extract" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.644417 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.647692 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.648107 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"galera-openstack-dockercfg-6qhpd" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.648176 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"kube-root-ca.crt" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.648192 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"openstack-config-data" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.648266 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"openstack-scripts" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.648691 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.648769 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"openshift-service-ca.crt" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.652471 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.653529 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.672189 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.678624 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.695050 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.798754 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-default\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.798811 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmkc6\" (UniqueName: \"kubernetes.io/projected/dde27b21-eedb-46d4-9cf5-99508535250e-kube-api-access-gmkc6\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.798839 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-operator-scripts\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.798929 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.798993 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-default\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799078 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-generated\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799111 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kolla-config\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799179 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-generated\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799273 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-kolla-config\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799331 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-generated\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799395 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799426 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-operator-scripts\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799461 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v7v5\" (UniqueName: \"kubernetes.io/projected/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kube-api-access-9v7v5\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799509 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tldhn\" (UniqueName: \"kubernetes.io/projected/225ba523-35eb-409e-8b11-9d1ec8994b38-kube-api-access-tldhn\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799595 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799631 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-default\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799709 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-kolla-config\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.799753 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-operator-scripts\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.900831 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-default\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.900889 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-kolla-config\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.900913 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-operator-scripts\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.900934 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-default\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.900954 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmkc6\" (UniqueName: \"kubernetes.io/projected/dde27b21-eedb-46d4-9cf5-99508535250e-kube-api-access-gmkc6\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.900969 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-operator-scripts\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.900986 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901004 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-default\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901021 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-generated\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901035 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kolla-config\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901055 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-generated\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901074 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-kolla-config\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901092 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-generated\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901109 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901124 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-operator-scripts\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901142 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v7v5\" (UniqueName: \"kubernetes.io/projected/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kube-api-access-9v7v5\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901160 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tldhn\" (UniqueName: \"kubernetes.io/projected/225ba523-35eb-409e-8b11-9d1ec8994b38-kube-api-access-tldhn\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901182 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901523 4718 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") device mount path \"/mnt/openstack/pv08\"" pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901918 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-kolla-config\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.901979 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-default\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.902175 4718 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") device mount path \"/mnt/openstack/pv05\"" pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.902485 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kolla-config\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.902548 4718 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") device mount path \"/mnt/openstack/pv06\"" pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.902703 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-generated\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.902952 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-generated\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.903080 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-generated\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.903891 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-operator-scripts\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.903909 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-operator-scripts\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.903995 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-default\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.904595 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-operator-scripts\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.905842 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-kolla-config\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.906023 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-default\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.922508 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.929017 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.929217 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.931837 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v7v5\" (UniqueName: \"kubernetes.io/projected/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kube-api-access-9v7v5\") pod \"openstack-galera-2\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.932445 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmkc6\" (UniqueName: \"kubernetes.io/projected/dde27b21-eedb-46d4-9cf5-99508535250e-kube-api-access-gmkc6\") pod \"openstack-galera-1\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.934784 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tldhn\" (UniqueName: \"kubernetes.io/projected/225ba523-35eb-409e-8b11-9d1ec8994b38-kube-api-access-tldhn\") pod \"openstack-galera-0\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:36 crc kubenswrapper[4718]: I1206 04:22:36.974062 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.000539 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.027890 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.095927 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc"] Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.096834 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.101845 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-czmtf" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.102036 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.113676 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc"] Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.207807 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-webhook-cert\") pod \"infra-operator-controller-manager-6df9d48c5f-fxqxc\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.207870 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-apiservice-cert\") pod \"infra-operator-controller-manager-6df9d48c5f-fxqxc\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.207917 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr5h8\" (UniqueName: \"kubernetes.io/projected/c51916d5-75f4-4435-b364-fc181ad7d21f-kube-api-access-mr5h8\") pod \"infra-operator-controller-manager-6df9d48c5f-fxqxc\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.309910 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr5h8\" (UniqueName: \"kubernetes.io/projected/c51916d5-75f4-4435-b364-fc181ad7d21f-kube-api-access-mr5h8\") pod \"infra-operator-controller-manager-6df9d48c5f-fxqxc\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.309995 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-webhook-cert\") pod \"infra-operator-controller-manager-6df9d48c5f-fxqxc\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.310042 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-apiservice-cert\") pod \"infra-operator-controller-manager-6df9d48c5f-fxqxc\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.315659 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-webhook-cert\") pod \"infra-operator-controller-manager-6df9d48c5f-fxqxc\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.315884 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-apiservice-cert\") pod \"infra-operator-controller-manager-6df9d48c5f-fxqxc\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.330379 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr5h8\" (UniqueName: \"kubernetes.io/projected/c51916d5-75f4-4435-b364-fc181ad7d21f-kube-api-access-mr5h8\") pod \"infra-operator-controller-manager-6df9d48c5f-fxqxc\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.422933 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.428698 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.547950 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0","Type":"ContainerStarted","Data":"5624a19eb6e41bb9a6603f207f51818574de46d40a0a4e730741ddd42629babf"} Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.568131 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Dec 06 04:22:37 crc kubenswrapper[4718]: W1206 04:22:37.573190 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddde27b21_eedb_46d4_9cf5_99508535250e.slice/crio-ffa83ff9075b47cfa5bc94b654cf3190a65a8f3525ecf1ff25736b77e5a79a0f WatchSource:0}: Error finding container ffa83ff9075b47cfa5bc94b654cf3190a65a8f3525ecf1ff25736b77e5a79a0f: Status 404 returned error can't find the container with id ffa83ff9075b47cfa5bc94b654cf3190a65a8f3525ecf1ff25736b77e5a79a0f Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.580086 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Dec 06 04:22:37 crc kubenswrapper[4718]: I1206 04:22:37.610366 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc"] Dec 06 04:22:37 crc kubenswrapper[4718]: W1206 04:22:37.618791 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc51916d5_75f4_4435_b364_fc181ad7d21f.slice/crio-db321dc8c9ada02785910a9602caf1e5bbceca1541d7ea6082d2dd721f571029 WatchSource:0}: Error finding container db321dc8c9ada02785910a9602caf1e5bbceca1541d7ea6082d2dd721f571029: Status 404 returned error can't find the container with id db321dc8c9ada02785910a9602caf1e5bbceca1541d7ea6082d2dd721f571029 Dec 06 04:22:38 crc kubenswrapper[4718]: I1206 04:22:38.555003 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"dde27b21-eedb-46d4-9cf5-99508535250e","Type":"ContainerStarted","Data":"ffa83ff9075b47cfa5bc94b654cf3190a65a8f3525ecf1ff25736b77e5a79a0f"} Dec 06 04:22:38 crc kubenswrapper[4718]: I1206 04:22:38.562868 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"225ba523-35eb-409e-8b11-9d1ec8994b38","Type":"ContainerStarted","Data":"2d587eae184b515ac35c3adcec3b88de68dcdcb1480c91ac4bd7c43496f26dd1"} Dec 06 04:22:38 crc kubenswrapper[4718]: I1206 04:22:38.564924 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" event={"ID":"c51916d5-75f4-4435-b364-fc181ad7d21f","Type":"ContainerStarted","Data":"db321dc8c9ada02785910a9602caf1e5bbceca1541d7ea6082d2dd721f571029"} Dec 06 04:22:47 crc kubenswrapper[4718]: I1206 04:22:47.658949 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" event={"ID":"c51916d5-75f4-4435-b364-fc181ad7d21f","Type":"ContainerStarted","Data":"60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90"} Dec 06 04:22:47 crc kubenswrapper[4718]: I1206 04:22:47.659515 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:47 crc kubenswrapper[4718]: I1206 04:22:47.664385 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0","Type":"ContainerStarted","Data":"6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6"} Dec 06 04:22:47 crc kubenswrapper[4718]: I1206 04:22:47.666065 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"dde27b21-eedb-46d4-9cf5-99508535250e","Type":"ContainerStarted","Data":"246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d"} Dec 06 04:22:47 crc kubenswrapper[4718]: I1206 04:22:47.667714 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"225ba523-35eb-409e-8b11-9d1ec8994b38","Type":"ContainerStarted","Data":"ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13"} Dec 06 04:22:47 crc kubenswrapper[4718]: I1206 04:22:47.708551 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" podStartSLOduration=1.055188996 podStartE2EDuration="10.708536034s" podCreationTimestamp="2025-12-06 04:22:37 +0000 UTC" firstStartedPulling="2025-12-06 04:22:37.622503667 +0000 UTC m=+946.628208828" lastFinishedPulling="2025-12-06 04:22:47.275850705 +0000 UTC m=+956.281555866" observedRunningTime="2025-12-06 04:22:47.684433086 +0000 UTC m=+956.690138257" watchObservedRunningTime="2025-12-06 04:22:47.708536034 +0000 UTC m=+956.714241185" Dec 06 04:22:51 crc kubenswrapper[4718]: I1206 04:22:51.697431 4718 generic.go:334] "Generic (PLEG): container finished" podID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" containerID="6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6" exitCode=0 Dec 06 04:22:51 crc kubenswrapper[4718]: I1206 04:22:51.697472 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0","Type":"ContainerDied","Data":"6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6"} Dec 06 04:22:51 crc kubenswrapper[4718]: I1206 04:22:51.699875 4718 generic.go:334] "Generic (PLEG): container finished" podID="dde27b21-eedb-46d4-9cf5-99508535250e" containerID="246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d" exitCode=0 Dec 06 04:22:51 crc kubenswrapper[4718]: I1206 04:22:51.699961 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"dde27b21-eedb-46d4-9cf5-99508535250e","Type":"ContainerDied","Data":"246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d"} Dec 06 04:22:51 crc kubenswrapper[4718]: I1206 04:22:51.702725 4718 generic.go:334] "Generic (PLEG): container finished" podID="225ba523-35eb-409e-8b11-9d1ec8994b38" containerID="ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13" exitCode=0 Dec 06 04:22:51 crc kubenswrapper[4718]: I1206 04:22:51.702770 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"225ba523-35eb-409e-8b11-9d1ec8994b38","Type":"ContainerDied","Data":"ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13"} Dec 06 04:22:52 crc kubenswrapper[4718]: I1206 04:22:52.709771 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0","Type":"ContainerStarted","Data":"a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b"} Dec 06 04:22:52 crc kubenswrapper[4718]: I1206 04:22:52.713287 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"dde27b21-eedb-46d4-9cf5-99508535250e","Type":"ContainerStarted","Data":"bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285"} Dec 06 04:22:52 crc kubenswrapper[4718]: I1206 04:22:52.715330 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"225ba523-35eb-409e-8b11-9d1ec8994b38","Type":"ContainerStarted","Data":"49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282"} Dec 06 04:22:52 crc kubenswrapper[4718]: I1206 04:22:52.726990 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/openstack-galera-2" podStartSLOduration=7.794003394 podStartE2EDuration="17.726972185s" podCreationTimestamp="2025-12-06 04:22:35 +0000 UTC" firstStartedPulling="2025-12-06 04:22:37.43763493 +0000 UTC m=+946.443340091" lastFinishedPulling="2025-12-06 04:22:47.370603721 +0000 UTC m=+956.376308882" observedRunningTime="2025-12-06 04:22:52.725132252 +0000 UTC m=+961.730837413" watchObservedRunningTime="2025-12-06 04:22:52.726972185 +0000 UTC m=+961.732677346" Dec 06 04:22:52 crc kubenswrapper[4718]: I1206 04:22:52.746597 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/openstack-galera-1" podStartSLOduration=7.903769354 podStartE2EDuration="17.746580533s" podCreationTimestamp="2025-12-06 04:22:35 +0000 UTC" firstStartedPulling="2025-12-06 04:22:37.578542743 +0000 UTC m=+946.584247904" lastFinishedPulling="2025-12-06 04:22:47.421353922 +0000 UTC m=+956.427059083" observedRunningTime="2025-12-06 04:22:52.745499371 +0000 UTC m=+961.751204532" watchObservedRunningTime="2025-12-06 04:22:52.746580533 +0000 UTC m=+961.752285694" Dec 06 04:22:52 crc kubenswrapper[4718]: I1206 04:22:52.767199 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/openstack-galera-0" podStartSLOduration=7.989674655 podStartE2EDuration="17.76717604s" podCreationTimestamp="2025-12-06 04:22:35 +0000 UTC" firstStartedPulling="2025-12-06 04:22:37.592599211 +0000 UTC m=+946.598304372" lastFinishedPulling="2025-12-06 04:22:47.370100596 +0000 UTC m=+956.375805757" observedRunningTime="2025-12-06 04:22:52.762411411 +0000 UTC m=+961.768116602" watchObservedRunningTime="2025-12-06 04:22:52.76717604 +0000 UTC m=+961.772881221" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.043502 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-222f4"] Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.045251 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.058953 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-222f4"] Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.111301 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-catalog-content\") pod \"community-operators-222f4\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.111395 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2vf8\" (UniqueName: \"kubernetes.io/projected/49458d1c-8d14-41ae-b52e-a3f21b29545d-kube-api-access-t2vf8\") pod \"community-operators-222f4\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.111417 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-utilities\") pod \"community-operators-222f4\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.212738 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2vf8\" (UniqueName: \"kubernetes.io/projected/49458d1c-8d14-41ae-b52e-a3f21b29545d-kube-api-access-t2vf8\") pod \"community-operators-222f4\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.212781 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-utilities\") pod \"community-operators-222f4\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.212854 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-catalog-content\") pod \"community-operators-222f4\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.213252 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-catalog-content\") pod \"community-operators-222f4\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.213362 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-utilities\") pod \"community-operators-222f4\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.232999 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2vf8\" (UniqueName: \"kubernetes.io/projected/49458d1c-8d14-41ae-b52e-a3f21b29545d-kube-api-access-t2vf8\") pod \"community-operators-222f4\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.365479 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-222f4" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.443713 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q2q7g"] Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.445088 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.509697 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2q7g"] Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.516715 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tspgj\" (UniqueName: \"kubernetes.io/projected/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-kube-api-access-tspgj\") pod \"redhat-marketplace-q2q7g\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.516779 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-utilities\") pod \"redhat-marketplace-q2q7g\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.516812 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-catalog-content\") pod \"redhat-marketplace-q2q7g\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.618495 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tspgj\" (UniqueName: \"kubernetes.io/projected/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-kube-api-access-tspgj\") pod \"redhat-marketplace-q2q7g\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.618559 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-utilities\") pod \"redhat-marketplace-q2q7g\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.618597 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-catalog-content\") pod \"redhat-marketplace-q2q7g\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.619092 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-catalog-content\") pod \"redhat-marketplace-q2q7g\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.619624 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-utilities\") pod \"redhat-marketplace-q2q7g\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.640944 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tspgj\" (UniqueName: \"kubernetes.io/projected/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-kube-api-access-tspgj\") pod \"redhat-marketplace-q2q7g\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.758083 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.865220 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-222f4"] Dec 06 04:22:56 crc kubenswrapper[4718]: W1206 04:22:56.875944 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49458d1c_8d14_41ae_b52e_a3f21b29545d.slice/crio-018d5803c6ee568191b7df910d52b6ccbd7ebd0003fa62afad7a9d4657c569b7 WatchSource:0}: Error finding container 018d5803c6ee568191b7df910d52b6ccbd7ebd0003fa62afad7a9d4657c569b7: Status 404 returned error can't find the container with id 018d5803c6ee568191b7df910d52b6ccbd7ebd0003fa62afad7a9d4657c569b7 Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.969763 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2q7g"] Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.974572 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:56 crc kubenswrapper[4718]: I1206 04:22:56.974785 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.000627 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.000659 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.029180 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.029223 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.427333 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.744615 4718 generic.go:334] "Generic (PLEG): container finished" podID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerID="12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4" exitCode=0 Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.744686 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-222f4" event={"ID":"49458d1c-8d14-41ae-b52e-a3f21b29545d","Type":"ContainerDied","Data":"12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4"} Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.744920 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-222f4" event={"ID":"49458d1c-8d14-41ae-b52e-a3f21b29545d","Type":"ContainerStarted","Data":"018d5803c6ee568191b7df910d52b6ccbd7ebd0003fa62afad7a9d4657c569b7"} Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.746876 4718 generic.go:334] "Generic (PLEG): container finished" podID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerID="40810f40da1888a95ac74ce637a3d49f0f516b00bdad382cdf126507708fd72c" exitCode=0 Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.747043 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2q7g" event={"ID":"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b","Type":"ContainerDied","Data":"40810f40da1888a95ac74ce637a3d49f0f516b00bdad382cdf126507708fd72c"} Dec 06 04:22:57 crc kubenswrapper[4718]: I1206 04:22:57.747094 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2q7g" event={"ID":"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b","Type":"ContainerStarted","Data":"ffb5da62627b220ca4daafc3b3999c3748f067b5691ac9987e797670a7dd5604"} Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.732814 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/memcached-0"] Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.733944 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.735773 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"memcached-memcached-dockercfg-4jm5l" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.736363 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"memcached-config-data" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.744143 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.855690 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-config-data\") pod \"memcached-0\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.855901 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-kolla-config\") pod \"memcached-0\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.855969 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f78dn\" (UniqueName: \"kubernetes.io/projected/f1e994e7-dee9-487f-aab6-aa711e080dcc-kube-api-access-f78dn\") pod \"memcached-0\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.957171 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-kolla-config\") pod \"memcached-0\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.957241 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f78dn\" (UniqueName: \"kubernetes.io/projected/f1e994e7-dee9-487f-aab6-aa711e080dcc-kube-api-access-f78dn\") pod \"memcached-0\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.957285 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-config-data\") pod \"memcached-0\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.958260 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-kolla-config\") pod \"memcached-0\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.958301 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-config-data\") pod \"memcached-0\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:58 crc kubenswrapper[4718]: I1206 04:22:58.979818 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f78dn\" (UniqueName: \"kubernetes.io/projected/f1e994e7-dee9-487f-aab6-aa711e080dcc-kube-api-access-f78dn\") pod \"memcached-0\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " pod="keystone-kuttl-tests/memcached-0" Dec 06 04:22:59 crc kubenswrapper[4718]: I1206 04:22:59.053811 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/memcached-0" Dec 06 04:23:02 crc kubenswrapper[4718]: I1206 04:23:02.182482 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Dec 06 04:23:02 crc kubenswrapper[4718]: I1206 04:23:02.206437 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:23:02 crc kubenswrapper[4718]: I1206 04:23:02.293254 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:23:02 crc kubenswrapper[4718]: I1206 04:23:02.779620 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/memcached-0" event={"ID":"f1e994e7-dee9-487f-aab6-aa711e080dcc","Type":"ContainerStarted","Data":"16c680d1112eb85b4abea4a597a860cb92fb4ad16e4e09e99840facfc45f4234"} Dec 06 04:23:02 crc kubenswrapper[4718]: I1206 04:23:02.781974 4718 generic.go:334] "Generic (PLEG): container finished" podID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerID="76be76c91e18c952087f5fbff0363df8d8c6d267261e9008783b1d4c5c36a9b0" exitCode=0 Dec 06 04:23:02 crc kubenswrapper[4718]: I1206 04:23:02.782078 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2q7g" event={"ID":"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b","Type":"ContainerDied","Data":"76be76c91e18c952087f5fbff0363df8d8c6d267261e9008783b1d4c5c36a9b0"} Dec 06 04:23:02 crc kubenswrapper[4718]: I1206 04:23:02.784329 4718 generic.go:334] "Generic (PLEG): container finished" podID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerID="526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c" exitCode=0 Dec 06 04:23:02 crc kubenswrapper[4718]: I1206 04:23:02.784405 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-222f4" event={"ID":"49458d1c-8d14-41ae-b52e-a3f21b29545d","Type":"ContainerDied","Data":"526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c"} Dec 06 04:23:03 crc kubenswrapper[4718]: I1206 04:23:03.032377 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-9w2ld"] Dec 06 04:23:03 crc kubenswrapper[4718]: I1206 04:23:03.033436 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:23:03 crc kubenswrapper[4718]: I1206 04:23:03.040124 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-pvmjt" Dec 06 04:23:03 crc kubenswrapper[4718]: I1206 04:23:03.050715 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-9w2ld"] Dec 06 04:23:03 crc kubenswrapper[4718]: I1206 04:23:03.118218 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ls96\" (UniqueName: \"kubernetes.io/projected/720bb2e9-e779-4b59-ac02-6f420b4dda74-kube-api-access-5ls96\") pod \"rabbitmq-cluster-operator-index-9w2ld\" (UID: \"720bb2e9-e779-4b59-ac02-6f420b4dda74\") " pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:23:03 crc kubenswrapper[4718]: I1206 04:23:03.220154 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ls96\" (UniqueName: \"kubernetes.io/projected/720bb2e9-e779-4b59-ac02-6f420b4dda74-kube-api-access-5ls96\") pod \"rabbitmq-cluster-operator-index-9w2ld\" (UID: \"720bb2e9-e779-4b59-ac02-6f420b4dda74\") " pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:23:03 crc kubenswrapper[4718]: I1206 04:23:03.259763 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ls96\" (UniqueName: \"kubernetes.io/projected/720bb2e9-e779-4b59-ac02-6f420b4dda74-kube-api-access-5ls96\") pod \"rabbitmq-cluster-operator-index-9w2ld\" (UID: \"720bb2e9-e779-4b59-ac02-6f420b4dda74\") " pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:23:03 crc kubenswrapper[4718]: I1206 04:23:03.349464 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:23:03 crc kubenswrapper[4718]: I1206 04:23:03.888127 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-9w2ld"] Dec 06 04:23:04 crc kubenswrapper[4718]: I1206 04:23:04.798834 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-222f4" event={"ID":"49458d1c-8d14-41ae-b52e-a3f21b29545d","Type":"ContainerStarted","Data":"fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c"} Dec 06 04:23:04 crc kubenswrapper[4718]: I1206 04:23:04.799965 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" event={"ID":"720bb2e9-e779-4b59-ac02-6f420b4dda74","Type":"ContainerStarted","Data":"f466930d7eb850b6ae1403b1fd5b84bc9fac81b82edb38f70ae8148ce463d43f"} Dec 06 04:23:04 crc kubenswrapper[4718]: I1206 04:23:04.829658 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-222f4" podStartSLOduration=1.94107837 podStartE2EDuration="8.829645312s" podCreationTimestamp="2025-12-06 04:22:56 +0000 UTC" firstStartedPulling="2025-12-06 04:22:57.746197399 +0000 UTC m=+966.751902560" lastFinishedPulling="2025-12-06 04:23:04.634764341 +0000 UTC m=+973.640469502" observedRunningTime="2025-12-06 04:23:04.825541972 +0000 UTC m=+973.831247133" watchObservedRunningTime="2025-12-06 04:23:04.829645312 +0000 UTC m=+973.835350473" Dec 06 04:23:05 crc kubenswrapper[4718]: I1206 04:23:05.810971 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2q7g" event={"ID":"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b","Type":"ContainerStarted","Data":"c0c59d324f3aba5902397c47d8a432896e91eecd3f79106f3ddaba3ed397809d"} Dec 06 04:23:05 crc kubenswrapper[4718]: I1206 04:23:05.835653 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q2q7g" podStartSLOduration=2.8954992969999997 podStartE2EDuration="9.835631802s" podCreationTimestamp="2025-12-06 04:22:56 +0000 UTC" firstStartedPulling="2025-12-06 04:22:57.748410813 +0000 UTC m=+966.754115974" lastFinishedPulling="2025-12-06 04:23:04.688543318 +0000 UTC m=+973.694248479" observedRunningTime="2025-12-06 04:23:05.827971102 +0000 UTC m=+974.833676263" watchObservedRunningTime="2025-12-06 04:23:05.835631802 +0000 UTC m=+974.841336963" Dec 06 04:23:06 crc kubenswrapper[4718]: I1206 04:23:06.366505 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-222f4" Dec 06 04:23:06 crc kubenswrapper[4718]: I1206 04:23:06.366704 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-222f4" Dec 06 04:23:06 crc kubenswrapper[4718]: I1206 04:23:06.758472 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:23:06 crc kubenswrapper[4718]: I1206 04:23:06.758673 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:23:06 crc kubenswrapper[4718]: I1206 04:23:06.808604 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:23:07 crc kubenswrapper[4718]: I1206 04:23:07.100343 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="keystone-kuttl-tests/openstack-galera-2" podUID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" containerName="galera" probeResult="failure" output=< Dec 06 04:23:07 crc kubenswrapper[4718]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Dec 06 04:23:07 crc kubenswrapper[4718]: > Dec 06 04:23:07 crc kubenswrapper[4718]: I1206 04:23:07.421213 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-222f4" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerName="registry-server" probeResult="failure" output=< Dec 06 04:23:07 crc kubenswrapper[4718]: timeout: failed to connect service ":50051" within 1s Dec 06 04:23:07 crc kubenswrapper[4718]: > Dec 06 04:23:08 crc kubenswrapper[4718]: I1206 04:23:08.835377 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tt249"] Dec 06 04:23:08 crc kubenswrapper[4718]: I1206 04:23:08.837291 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:08 crc kubenswrapper[4718]: I1206 04:23:08.847847 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt249"] Dec 06 04:23:08 crc kubenswrapper[4718]: I1206 04:23:08.913100 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzrbw\" (UniqueName: \"kubernetes.io/projected/2a5624c1-2492-44d1-8e2c-ea19f74d275e-kube-api-access-jzrbw\") pod \"certified-operators-tt249\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:08 crc kubenswrapper[4718]: I1206 04:23:08.913153 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-catalog-content\") pod \"certified-operators-tt249\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:08 crc kubenswrapper[4718]: I1206 04:23:08.913319 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-utilities\") pod \"certified-operators-tt249\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:09 crc kubenswrapper[4718]: I1206 04:23:09.014075 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzrbw\" (UniqueName: \"kubernetes.io/projected/2a5624c1-2492-44d1-8e2c-ea19f74d275e-kube-api-access-jzrbw\") pod \"certified-operators-tt249\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:09 crc kubenswrapper[4718]: I1206 04:23:09.014129 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-catalog-content\") pod \"certified-operators-tt249\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:09 crc kubenswrapper[4718]: I1206 04:23:09.014164 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-utilities\") pod \"certified-operators-tt249\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:09 crc kubenswrapper[4718]: I1206 04:23:09.014917 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-catalog-content\") pod \"certified-operators-tt249\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:09 crc kubenswrapper[4718]: I1206 04:23:09.015163 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-utilities\") pod \"certified-operators-tt249\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:09 crc kubenswrapper[4718]: I1206 04:23:09.062089 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzrbw\" (UniqueName: \"kubernetes.io/projected/2a5624c1-2492-44d1-8e2c-ea19f74d275e-kube-api-access-jzrbw\") pod \"certified-operators-tt249\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:09 crc kubenswrapper[4718]: I1206 04:23:09.222883 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:10 crc kubenswrapper[4718]: I1206 04:23:10.339939 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt249"] Dec 06 04:23:11 crc kubenswrapper[4718]: W1206 04:23:11.053624 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a5624c1_2492_44d1_8e2c_ea19f74d275e.slice/crio-ad9a44fabc8a5ec623eb62994eddbabea761c1261c2f48fcefa7a1a7ee87cf52 WatchSource:0}: Error finding container ad9a44fabc8a5ec623eb62994eddbabea761c1261c2f48fcefa7a1a7ee87cf52: Status 404 returned error can't find the container with id ad9a44fabc8a5ec623eb62994eddbabea761c1261c2f48fcefa7a1a7ee87cf52 Dec 06 04:23:11 crc kubenswrapper[4718]: I1206 04:23:11.856499 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" event={"ID":"720bb2e9-e779-4b59-ac02-6f420b4dda74","Type":"ContainerStarted","Data":"31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3"} Dec 06 04:23:11 crc kubenswrapper[4718]: I1206 04:23:11.860524 4718 generic.go:334] "Generic (PLEG): container finished" podID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerID="9cd2da3f844c770f569463cd077514625475d0eacdecff143dc18bfde9601a9f" exitCode=0 Dec 06 04:23:11 crc kubenswrapper[4718]: I1206 04:23:11.860641 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt249" event={"ID":"2a5624c1-2492-44d1-8e2c-ea19f74d275e","Type":"ContainerDied","Data":"9cd2da3f844c770f569463cd077514625475d0eacdecff143dc18bfde9601a9f"} Dec 06 04:23:11 crc kubenswrapper[4718]: I1206 04:23:11.860713 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt249" event={"ID":"2a5624c1-2492-44d1-8e2c-ea19f74d275e","Type":"ContainerStarted","Data":"ad9a44fabc8a5ec623eb62994eddbabea761c1261c2f48fcefa7a1a7ee87cf52"} Dec 06 04:23:11 crc kubenswrapper[4718]: I1206 04:23:11.865543 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/memcached-0" event={"ID":"f1e994e7-dee9-487f-aab6-aa711e080dcc","Type":"ContainerStarted","Data":"bdea561e2e9b28034932b4af606975c07a361078d7406c942c644d036717d304"} Dec 06 04:23:11 crc kubenswrapper[4718]: I1206 04:23:11.865723 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/memcached-0" Dec 06 04:23:11 crc kubenswrapper[4718]: I1206 04:23:11.875182 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" podStartSLOduration=1.658791918 podStartE2EDuration="8.875164585s" podCreationTimestamp="2025-12-06 04:23:03 +0000 UTC" firstStartedPulling="2025-12-06 04:23:03.895636618 +0000 UTC m=+972.901341779" lastFinishedPulling="2025-12-06 04:23:11.112009285 +0000 UTC m=+980.117714446" observedRunningTime="2025-12-06 04:23:11.874807193 +0000 UTC m=+980.880512354" watchObservedRunningTime="2025-12-06 04:23:11.875164585 +0000 UTC m=+980.880869736" Dec 06 04:23:11 crc kubenswrapper[4718]: I1206 04:23:11.898930 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/memcached-0" podStartSLOduration=7.179074673 podStartE2EDuration="13.898899311s" podCreationTimestamp="2025-12-06 04:22:58 +0000 UTC" firstStartedPulling="2025-12-06 04:23:02.192558452 +0000 UTC m=+971.198263613" lastFinishedPulling="2025-12-06 04:23:08.91238309 +0000 UTC m=+977.918088251" observedRunningTime="2025-12-06 04:23:11.898593381 +0000 UTC m=+980.904298542" watchObservedRunningTime="2025-12-06 04:23:11.898899311 +0000 UTC m=+980.904604472" Dec 06 04:23:13 crc kubenswrapper[4718]: I1206 04:23:13.350074 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:23:13 crc kubenswrapper[4718]: I1206 04:23:13.350505 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:23:13 crc kubenswrapper[4718]: I1206 04:23:13.380465 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:23:13 crc kubenswrapper[4718]: I1206 04:23:13.880361 4718 generic.go:334] "Generic (PLEG): container finished" podID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerID="c463ab0932fb381723f3b58f4217359e95ce5e14535877c1c9803046c14c6e79" exitCode=0 Dec 06 04:23:13 crc kubenswrapper[4718]: I1206 04:23:13.880783 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt249" event={"ID":"2a5624c1-2492-44d1-8e2c-ea19f74d275e","Type":"ContainerDied","Data":"c463ab0932fb381723f3b58f4217359e95ce5e14535877c1c9803046c14c6e79"} Dec 06 04:23:14 crc kubenswrapper[4718]: I1206 04:23:14.899508 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt249" event={"ID":"2a5624c1-2492-44d1-8e2c-ea19f74d275e","Type":"ContainerStarted","Data":"67e9a917a0d8a668433cf542eb7fb768eb30ac5c683452fa895b345a77629c60"} Dec 06 04:23:14 crc kubenswrapper[4718]: I1206 04:23:14.927851 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tt249" podStartSLOduration=4.482037644 podStartE2EDuration="6.927832374s" podCreationTimestamp="2025-12-06 04:23:08 +0000 UTC" firstStartedPulling="2025-12-06 04:23:11.862907878 +0000 UTC m=+980.868613039" lastFinishedPulling="2025-12-06 04:23:14.308702598 +0000 UTC m=+983.314407769" observedRunningTime="2025-12-06 04:23:14.92417887 +0000 UTC m=+983.929884031" watchObservedRunningTime="2025-12-06 04:23:14.927832374 +0000 UTC m=+983.933537545" Dec 06 04:23:15 crc kubenswrapper[4718]: I1206 04:23:15.131321 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:23:15 crc kubenswrapper[4718]: I1206 04:23:15.222445 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:23:16 crc kubenswrapper[4718]: I1206 04:23:16.414296 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-222f4" Dec 06 04:23:16 crc kubenswrapper[4718]: I1206 04:23:16.450416 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-222f4" Dec 06 04:23:16 crc kubenswrapper[4718]: I1206 04:23:16.892377 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:23:17 crc kubenswrapper[4718]: I1206 04:23:17.792886 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:23:17 crc kubenswrapper[4718]: I1206 04:23:17.899003 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:23:19 crc kubenswrapper[4718]: I1206 04:23:19.055272 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/memcached-0" Dec 06 04:23:19 crc kubenswrapper[4718]: I1206 04:23:19.223759 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:19 crc kubenswrapper[4718]: I1206 04:23:19.224179 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:19 crc kubenswrapper[4718]: I1206 04:23:19.300057 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:19 crc kubenswrapper[4718]: I1206 04:23:19.999098 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:20 crc kubenswrapper[4718]: I1206 04:23:20.431342 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-222f4"] Dec 06 04:23:20 crc kubenswrapper[4718]: I1206 04:23:20.431695 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-222f4" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerName="registry-server" containerID="cri-o://fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c" gracePeriod=2 Dec 06 04:23:20 crc kubenswrapper[4718]: I1206 04:23:20.828467 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2q7g"] Dec 06 04:23:20 crc kubenswrapper[4718]: I1206 04:23:20.828749 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q2q7g" podUID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerName="registry-server" containerID="cri-o://c0c59d324f3aba5902397c47d8a432896e91eecd3f79106f3ddaba3ed397809d" gracePeriod=2 Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.923106 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-222f4" Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.942682 4718 generic.go:334] "Generic (PLEG): container finished" podID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerID="c0c59d324f3aba5902397c47d8a432896e91eecd3f79106f3ddaba3ed397809d" exitCode=0 Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.942742 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2q7g" event={"ID":"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b","Type":"ContainerDied","Data":"c0c59d324f3aba5902397c47d8a432896e91eecd3f79106f3ddaba3ed397809d"} Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.944512 4718 generic.go:334] "Generic (PLEG): container finished" podID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerID="fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c" exitCode=0 Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.944545 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-222f4" event={"ID":"49458d1c-8d14-41ae-b52e-a3f21b29545d","Type":"ContainerDied","Data":"fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c"} Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.944564 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-222f4" event={"ID":"49458d1c-8d14-41ae-b52e-a3f21b29545d","Type":"ContainerDied","Data":"018d5803c6ee568191b7df910d52b6ccbd7ebd0003fa62afad7a9d4657c569b7"} Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.944586 4718 scope.go:117] "RemoveContainer" containerID="fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c" Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.944717 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-222f4" Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.963150 4718 scope.go:117] "RemoveContainer" containerID="526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c" Dec 06 04:23:21 crc kubenswrapper[4718]: I1206 04:23:21.980954 4718 scope.go:117] "RemoveContainer" containerID="12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.007541 4718 scope.go:117] "RemoveContainer" containerID="fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c" Dec 06 04:23:22 crc kubenswrapper[4718]: E1206 04:23:22.008023 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c\": container with ID starting with fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c not found: ID does not exist" containerID="fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.008068 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c"} err="failed to get container status \"fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c\": rpc error: code = NotFound desc = could not find container \"fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c\": container with ID starting with fc57486b8b892840d3319a4c473170c315f997f1619173d265c7bfd6a85a857c not found: ID does not exist" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.008094 4718 scope.go:117] "RemoveContainer" containerID="526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c" Dec 06 04:23:22 crc kubenswrapper[4718]: E1206 04:23:22.008662 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c\": container with ID starting with 526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c not found: ID does not exist" containerID="526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.008699 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c"} err="failed to get container status \"526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c\": rpc error: code = NotFound desc = could not find container \"526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c\": container with ID starting with 526bdb252c1d128c8da6f34bdb09c27a8712c039e78afcfcfc44f072f434385c not found: ID does not exist" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.008724 4718 scope.go:117] "RemoveContainer" containerID="12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4" Dec 06 04:23:22 crc kubenswrapper[4718]: E1206 04:23:22.009114 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4\": container with ID starting with 12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4 not found: ID does not exist" containerID="12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.009149 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4"} err="failed to get container status \"12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4\": rpc error: code = NotFound desc = could not find container \"12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4\": container with ID starting with 12a7acbc7501224016329b3d1d73227f56850a6bd69fd727c95ed5e856a89fd4 not found: ID does not exist" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.080016 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-utilities\") pod \"49458d1c-8d14-41ae-b52e-a3f21b29545d\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.080136 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-catalog-content\") pod \"49458d1c-8d14-41ae-b52e-a3f21b29545d\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.080260 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2vf8\" (UniqueName: \"kubernetes.io/projected/49458d1c-8d14-41ae-b52e-a3f21b29545d-kube-api-access-t2vf8\") pod \"49458d1c-8d14-41ae-b52e-a3f21b29545d\" (UID: \"49458d1c-8d14-41ae-b52e-a3f21b29545d\") " Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.081414 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-utilities" (OuterVolumeSpecName: "utilities") pod "49458d1c-8d14-41ae-b52e-a3f21b29545d" (UID: "49458d1c-8d14-41ae-b52e-a3f21b29545d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.086385 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49458d1c-8d14-41ae-b52e-a3f21b29545d-kube-api-access-t2vf8" (OuterVolumeSpecName: "kube-api-access-t2vf8") pod "49458d1c-8d14-41ae-b52e-a3f21b29545d" (UID: "49458d1c-8d14-41ae-b52e-a3f21b29545d"). InnerVolumeSpecName "kube-api-access-t2vf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.132192 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49458d1c-8d14-41ae-b52e-a3f21b29545d" (UID: "49458d1c-8d14-41ae-b52e-a3f21b29545d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.181908 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.181963 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49458d1c-8d14-41ae-b52e-a3f21b29545d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.181988 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2vf8\" (UniqueName: \"kubernetes.io/projected/49458d1c-8d14-41ae-b52e-a3f21b29545d-kube-api-access-t2vf8\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.263868 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.310771 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-222f4"] Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.314216 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-222f4"] Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.384962 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-catalog-content\") pod \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.385062 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tspgj\" (UniqueName: \"kubernetes.io/projected/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-kube-api-access-tspgj\") pod \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.385222 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-utilities\") pod \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\" (UID: \"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b\") " Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.390060 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-utilities" (OuterVolumeSpecName: "utilities") pod "b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" (UID: "b7223f0f-f30e-4d4b-91d2-f699c2f5e34b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.392181 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-kube-api-access-tspgj" (OuterVolumeSpecName: "kube-api-access-tspgj") pod "b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" (UID: "b7223f0f-f30e-4d4b-91d2-f699c2f5e34b"). InnerVolumeSpecName "kube-api-access-tspgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.412089 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" (UID: "b7223f0f-f30e-4d4b-91d2-f699c2f5e34b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.487150 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.487185 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tspgj\" (UniqueName: \"kubernetes.io/projected/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-kube-api-access-tspgj\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.487196 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.955432 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2q7g" event={"ID":"b7223f0f-f30e-4d4b-91d2-f699c2f5e34b","Type":"ContainerDied","Data":"ffb5da62627b220ca4daafc3b3999c3748f067b5691ac9987e797670a7dd5604"} Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.955542 4718 scope.go:117] "RemoveContainer" containerID="c0c59d324f3aba5902397c47d8a432896e91eecd3f79106f3ddaba3ed397809d" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.956718 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2q7g" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.980203 4718 scope.go:117] "RemoveContainer" containerID="76be76c91e18c952087f5fbff0363df8d8c6d267261e9008783b1d4c5c36a9b0" Dec 06 04:23:22 crc kubenswrapper[4718]: I1206 04:23:22.999838 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2q7g"] Dec 06 04:23:23 crc kubenswrapper[4718]: I1206 04:23:23.004180 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2q7g"] Dec 06 04:23:23 crc kubenswrapper[4718]: I1206 04:23:23.021463 4718 scope.go:117] "RemoveContainer" containerID="40810f40da1888a95ac74ce637a3d49f0f516b00bdad382cdf126507708fd72c" Dec 06 04:23:23 crc kubenswrapper[4718]: I1206 04:23:23.345150 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" path="/var/lib/kubelet/pods/49458d1c-8d14-41ae-b52e-a3f21b29545d/volumes" Dec 06 04:23:23 crc kubenswrapper[4718]: I1206 04:23:23.347296 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" path="/var/lib/kubelet/pods/b7223f0f-f30e-4d4b-91d2-f699c2f5e34b/volumes" Dec 06 04:23:23 crc kubenswrapper[4718]: I1206 04:23:23.399332 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:23:25 crc kubenswrapper[4718]: I1206 04:23:25.224144 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tt249"] Dec 06 04:23:25 crc kubenswrapper[4718]: I1206 04:23:25.225146 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tt249" podUID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerName="registry-server" containerID="cri-o://67e9a917a0d8a668433cf542eb7fb768eb30ac5c683452fa895b345a77629c60" gracePeriod=2 Dec 06 04:23:25 crc kubenswrapper[4718]: I1206 04:23:25.982791 4718 generic.go:334] "Generic (PLEG): container finished" podID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerID="67e9a917a0d8a668433cf542eb7fb768eb30ac5c683452fa895b345a77629c60" exitCode=0 Dec 06 04:23:25 crc kubenswrapper[4718]: I1206 04:23:25.982899 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt249" event={"ID":"2a5624c1-2492-44d1-8e2c-ea19f74d275e","Type":"ContainerDied","Data":"67e9a917a0d8a668433cf542eb7fb768eb30ac5c683452fa895b345a77629c60"} Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.247335 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.335319 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-catalog-content\") pod \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.335444 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-utilities\") pod \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.335588 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzrbw\" (UniqueName: \"kubernetes.io/projected/2a5624c1-2492-44d1-8e2c-ea19f74d275e-kube-api-access-jzrbw\") pod \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\" (UID: \"2a5624c1-2492-44d1-8e2c-ea19f74d275e\") " Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.336345 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-utilities" (OuterVolumeSpecName: "utilities") pod "2a5624c1-2492-44d1-8e2c-ea19f74d275e" (UID: "2a5624c1-2492-44d1-8e2c-ea19f74d275e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.339853 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a5624c1-2492-44d1-8e2c-ea19f74d275e-kube-api-access-jzrbw" (OuterVolumeSpecName: "kube-api-access-jzrbw") pod "2a5624c1-2492-44d1-8e2c-ea19f74d275e" (UID: "2a5624c1-2492-44d1-8e2c-ea19f74d275e"). InnerVolumeSpecName "kube-api-access-jzrbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.384434 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a5624c1-2492-44d1-8e2c-ea19f74d275e" (UID: "2a5624c1-2492-44d1-8e2c-ea19f74d275e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.438485 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.439515 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5624c1-2492-44d1-8e2c-ea19f74d275e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.439791 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzrbw\" (UniqueName: \"kubernetes.io/projected/2a5624c1-2492-44d1-8e2c-ea19f74d275e-kube-api-access-jzrbw\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.994943 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt249" event={"ID":"2a5624c1-2492-44d1-8e2c-ea19f74d275e","Type":"ContainerDied","Data":"ad9a44fabc8a5ec623eb62994eddbabea761c1261c2f48fcefa7a1a7ee87cf52"} Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.995006 4718 scope.go:117] "RemoveContainer" containerID="67e9a917a0d8a668433cf542eb7fb768eb30ac5c683452fa895b345a77629c60" Dec 06 04:23:26 crc kubenswrapper[4718]: I1206 04:23:26.995006 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt249" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.025439 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tt249"] Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.031992 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tt249"] Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.034338 4718 scope.go:117] "RemoveContainer" containerID="c463ab0932fb381723f3b58f4217359e95ce5e14535877c1c9803046c14c6e79" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.071813 4718 scope.go:117] "RemoveContainer" containerID="9cd2da3f844c770f569463cd077514625475d0eacdecff143dc18bfde9601a9f" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.346954 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" path="/var/lib/kubelet/pods/2a5624c1-2492-44d1-8e2c-ea19f74d275e/volumes" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.708073 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb"] Dec 06 04:23:27 crc kubenswrapper[4718]: E1206 04:23:27.708558 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerName="registry-server" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.708636 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerName="registry-server" Dec 06 04:23:27 crc kubenswrapper[4718]: E1206 04:23:27.708677 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerName="registry-server" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.708696 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerName="registry-server" Dec 06 04:23:27 crc kubenswrapper[4718]: E1206 04:23:27.708723 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerName="registry-server" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.708740 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerName="registry-server" Dec 06 04:23:27 crc kubenswrapper[4718]: E1206 04:23:27.708777 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerName="extract-utilities" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.708795 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerName="extract-utilities" Dec 06 04:23:27 crc kubenswrapper[4718]: E1206 04:23:27.708818 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerName="extract-utilities" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.708835 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerName="extract-utilities" Dec 06 04:23:27 crc kubenswrapper[4718]: E1206 04:23:27.708864 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerName="extract-content" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.708880 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerName="extract-content" Dec 06 04:23:27 crc kubenswrapper[4718]: E1206 04:23:27.708907 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerName="extract-utilities" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.708925 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerName="extract-utilities" Dec 06 04:23:27 crc kubenswrapper[4718]: E1206 04:23:27.708957 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerName="extract-content" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.708975 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerName="extract-content" Dec 06 04:23:27 crc kubenswrapper[4718]: E1206 04:23:27.709005 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerName="extract-content" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.709023 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerName="extract-content" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.709321 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7223f0f-f30e-4d4b-91d2-f699c2f5e34b" containerName="registry-server" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.709349 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a5624c1-2492-44d1-8e2c-ea19f74d275e" containerName="registry-server" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.709391 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="49458d1c-8d14-41ae-b52e-a3f21b29545d" containerName="registry-server" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.713589 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.716369 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-69frt" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.725302 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb"] Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.759039 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.759092 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.759151 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpsd5\" (UniqueName: \"kubernetes.io/projected/5f14df8d-3e33-475f-8e17-f507f1aaf429-kube-api-access-hpsd5\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.860100 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.860152 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.860217 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpsd5\" (UniqueName: \"kubernetes.io/projected/5f14df8d-3e33-475f-8e17-f507f1aaf429-kube-api-access-hpsd5\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.860788 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.860878 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.877313 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.877380 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:23:27 crc kubenswrapper[4718]: I1206 04:23:27.884683 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpsd5\" (UniqueName: \"kubernetes.io/projected/5f14df8d-3e33-475f-8e17-f507f1aaf429-kube-api-access-hpsd5\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:28 crc kubenswrapper[4718]: I1206 04:23:28.040965 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:28 crc kubenswrapper[4718]: I1206 04:23:28.489810 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb"] Dec 06 04:23:29 crc kubenswrapper[4718]: I1206 04:23:29.011157 4718 generic.go:334] "Generic (PLEG): container finished" podID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerID="6bfed44d0979e803bc0093f058cc12be1bdcf12fff577d2c92d5fbe36eadd256" exitCode=0 Dec 06 04:23:29 crc kubenswrapper[4718]: I1206 04:23:29.011256 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" event={"ID":"5f14df8d-3e33-475f-8e17-f507f1aaf429","Type":"ContainerDied","Data":"6bfed44d0979e803bc0093f058cc12be1bdcf12fff577d2c92d5fbe36eadd256"} Dec 06 04:23:29 crc kubenswrapper[4718]: I1206 04:23:29.011289 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" event={"ID":"5f14df8d-3e33-475f-8e17-f507f1aaf429","Type":"ContainerStarted","Data":"2fc2ffe48c4a3fc3f0ca4632caa24c276f117a5142a5cc1036d7622813bc7be3"} Dec 06 04:23:30 crc kubenswrapper[4718]: I1206 04:23:30.018505 4718 generic.go:334] "Generic (PLEG): container finished" podID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerID="bd262fbe949bba5ca68d8d315477d0ac3f1fd75b84cb448db0cd1a01602fc00a" exitCode=0 Dec 06 04:23:30 crc kubenswrapper[4718]: I1206 04:23:30.018591 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" event={"ID":"5f14df8d-3e33-475f-8e17-f507f1aaf429","Type":"ContainerDied","Data":"bd262fbe949bba5ca68d8d315477d0ac3f1fd75b84cb448db0cd1a01602fc00a"} Dec 06 04:23:31 crc kubenswrapper[4718]: I1206 04:23:31.028591 4718 generic.go:334] "Generic (PLEG): container finished" podID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerID="61b2860f6878a3eb87308c49873c973e01b93112967cfd54ab36e38fdcc98475" exitCode=0 Dec 06 04:23:31 crc kubenswrapper[4718]: I1206 04:23:31.028651 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" event={"ID":"5f14df8d-3e33-475f-8e17-f507f1aaf429","Type":"ContainerDied","Data":"61b2860f6878a3eb87308c49873c973e01b93112967cfd54ab36e38fdcc98475"} Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.360473 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.433648 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-util\") pod \"5f14df8d-3e33-475f-8e17-f507f1aaf429\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.434150 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-bundle\") pod \"5f14df8d-3e33-475f-8e17-f507f1aaf429\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.434328 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpsd5\" (UniqueName: \"kubernetes.io/projected/5f14df8d-3e33-475f-8e17-f507f1aaf429-kube-api-access-hpsd5\") pod \"5f14df8d-3e33-475f-8e17-f507f1aaf429\" (UID: \"5f14df8d-3e33-475f-8e17-f507f1aaf429\") " Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.434891 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-bundle" (OuterVolumeSpecName: "bundle") pod "5f14df8d-3e33-475f-8e17-f507f1aaf429" (UID: "5f14df8d-3e33-475f-8e17-f507f1aaf429"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.435226 4718 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.441760 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f14df8d-3e33-475f-8e17-f507f1aaf429-kube-api-access-hpsd5" (OuterVolumeSpecName: "kube-api-access-hpsd5") pod "5f14df8d-3e33-475f-8e17-f507f1aaf429" (UID: "5f14df8d-3e33-475f-8e17-f507f1aaf429"). InnerVolumeSpecName "kube-api-access-hpsd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.457712 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-util" (OuterVolumeSpecName: "util") pod "5f14df8d-3e33-475f-8e17-f507f1aaf429" (UID: "5f14df8d-3e33-475f-8e17-f507f1aaf429"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.537182 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpsd5\" (UniqueName: \"kubernetes.io/projected/5f14df8d-3e33-475f-8e17-f507f1aaf429-kube-api-access-hpsd5\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:32 crc kubenswrapper[4718]: I1206 04:23:32.537267 4718 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f14df8d-3e33-475f-8e17-f507f1aaf429-util\") on node \"crc\" DevicePath \"\"" Dec 06 04:23:33 crc kubenswrapper[4718]: I1206 04:23:33.043817 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" event={"ID":"5f14df8d-3e33-475f-8e17-f507f1aaf429","Type":"ContainerDied","Data":"2fc2ffe48c4a3fc3f0ca4632caa24c276f117a5142a5cc1036d7622813bc7be3"} Dec 06 04:23:33 crc kubenswrapper[4718]: I1206 04:23:33.043853 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fc2ffe48c4a3fc3f0ca4632caa24c276f117a5142a5cc1036d7622813bc7be3" Dec 06 04:23:33 crc kubenswrapper[4718]: I1206 04:23:33.043907 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.591916 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw"] Dec 06 04:23:45 crc kubenswrapper[4718]: E1206 04:23:45.593146 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerName="extract" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.593174 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerName="extract" Dec 06 04:23:45 crc kubenswrapper[4718]: E1206 04:23:45.593205 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerName="pull" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.593220 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerName="pull" Dec 06 04:23:45 crc kubenswrapper[4718]: E1206 04:23:45.593276 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerName="util" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.593296 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerName="util" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.593551 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f14df8d-3e33-475f-8e17-f507f1aaf429" containerName="extract" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.594408 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.599807 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-rgfk8" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.605954 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw"] Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.622552 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpvrb\" (UniqueName: \"kubernetes.io/projected/513e6338-5c23-4b84-96b8-bd15003f6595-kube-api-access-vpvrb\") pod \"rabbitmq-cluster-operator-779fc9694b-cb7pw\" (UID: \"513e6338-5c23-4b84-96b8-bd15003f6595\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.723558 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpvrb\" (UniqueName: \"kubernetes.io/projected/513e6338-5c23-4b84-96b8-bd15003f6595-kube-api-access-vpvrb\") pod \"rabbitmq-cluster-operator-779fc9694b-cb7pw\" (UID: \"513e6338-5c23-4b84-96b8-bd15003f6595\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.745694 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpvrb\" (UniqueName: \"kubernetes.io/projected/513e6338-5c23-4b84-96b8-bd15003f6595-kube-api-access-vpvrb\") pod \"rabbitmq-cluster-operator-779fc9694b-cb7pw\" (UID: \"513e6338-5c23-4b84-96b8-bd15003f6595\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" Dec 06 04:23:45 crc kubenswrapper[4718]: I1206 04:23:45.919967 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" Dec 06 04:23:46 crc kubenswrapper[4718]: I1206 04:23:46.150126 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw"] Dec 06 04:23:47 crc kubenswrapper[4718]: I1206 04:23:47.149310 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" event={"ID":"513e6338-5c23-4b84-96b8-bd15003f6595","Type":"ContainerStarted","Data":"f9e78616187c0b42d6d72501f4533a6633fdfce78667db7a8e145a5e88da3a3e"} Dec 06 04:23:49 crc kubenswrapper[4718]: I1206 04:23:49.162482 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" event={"ID":"513e6338-5c23-4b84-96b8-bd15003f6595","Type":"ContainerStarted","Data":"7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62"} Dec 06 04:23:49 crc kubenswrapper[4718]: I1206 04:23:49.183116 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" podStartSLOduration=1.485110891 podStartE2EDuration="4.18309405s" podCreationTimestamp="2025-12-06 04:23:45 +0000 UTC" firstStartedPulling="2025-12-06 04:23:46.164006901 +0000 UTC m=+1015.169712062" lastFinishedPulling="2025-12-06 04:23:48.86199004 +0000 UTC m=+1017.867695221" observedRunningTime="2025-12-06 04:23:49.178720271 +0000 UTC m=+1018.184425462" watchObservedRunningTime="2025-12-06 04:23:49.18309405 +0000 UTC m=+1018.188799251" Dec 06 04:23:51 crc kubenswrapper[4718]: I1206 04:23:51.898961 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Dec 06 04:23:51 crc kubenswrapper[4718]: I1206 04:23:51.900560 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:51 crc kubenswrapper[4718]: I1206 04:23:51.903580 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"rabbitmq-erlang-cookie" Dec 06 04:23:51 crc kubenswrapper[4718]: I1206 04:23:51.904055 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"rabbitmq-plugins-conf" Dec 06 04:23:51 crc kubenswrapper[4718]: I1206 04:23:51.904056 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"rabbitmq-default-user" Dec 06 04:23:51 crc kubenswrapper[4718]: I1206 04:23:51.913024 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"rabbitmq-server-conf" Dec 06 04:23:51 crc kubenswrapper[4718]: I1206 04:23:51.913249 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"rabbitmq-server-dockercfg-wqhqr" Dec 06 04:23:51 crc kubenswrapper[4718]: I1206 04:23:51.962003 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.012003 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.012051 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e9868af3-c211-4dde-9701-669e579f484b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.012093 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfcng\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-kube-api-access-hfcng\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.012116 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.012133 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e9868af3-c211-4dde-9701-669e579f484b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.012155 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e9868af3-c211-4dde-9701-669e579f484b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.012397 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.012494 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.113753 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.114493 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e9868af3-c211-4dde-9701-669e579f484b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.114573 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfcng\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-kube-api-access-hfcng\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.114668 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.114741 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e9868af3-c211-4dde-9701-669e579f484b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.114961 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e9868af3-c211-4dde-9701-669e579f484b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.115037 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.115927 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.116026 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.116384 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.118108 4718 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.118147 4718 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c6d9d48707b126fbec37391e863e6652330ea5eea3ee3dc01410ab8622cefa39/globalmount\"" pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.118159 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e9868af3-c211-4dde-9701-669e579f484b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.119994 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e9868af3-c211-4dde-9701-669e579f484b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.120649 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.129618 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e9868af3-c211-4dde-9701-669e579f484b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.139051 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfcng\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-kube-api-access-hfcng\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.152412 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\") pod \"rabbitmq-server-0\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.215565 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:23:52 crc kubenswrapper[4718]: I1206 04:23:52.698732 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Dec 06 04:23:52 crc kubenswrapper[4718]: W1206 04:23:52.708201 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9868af3_c211_4dde_9701_669e579f484b.slice/crio-339e2b442af7dbb894f958c84f1b16d4f17090248a4696daf5c7f88eb4332f14 WatchSource:0}: Error finding container 339e2b442af7dbb894f958c84f1b16d4f17090248a4696daf5c7f88eb4332f14: Status 404 returned error can't find the container with id 339e2b442af7dbb894f958c84f1b16d4f17090248a4696daf5c7f88eb4332f14 Dec 06 04:23:53 crc kubenswrapper[4718]: I1206 04:23:53.194970 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"e9868af3-c211-4dde-9701-669e579f484b","Type":"ContainerStarted","Data":"339e2b442af7dbb894f958c84f1b16d4f17090248a4696daf5c7f88eb4332f14"} Dec 06 04:23:53 crc kubenswrapper[4718]: I1206 04:23:53.647030 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-hql9d"] Dec 06 04:23:53 crc kubenswrapper[4718]: I1206 04:23:53.648555 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:23:53 crc kubenswrapper[4718]: I1206 04:23:53.650945 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-fvp2x" Dec 06 04:23:53 crc kubenswrapper[4718]: I1206 04:23:53.655311 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-hql9d"] Dec 06 04:23:53 crc kubenswrapper[4718]: I1206 04:23:53.838698 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq5w4\" (UniqueName: \"kubernetes.io/projected/147bc60c-8979-462d-bfc5-8460b289738d-kube-api-access-vq5w4\") pod \"keystone-operator-index-hql9d\" (UID: \"147bc60c-8979-462d-bfc5-8460b289738d\") " pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:23:53 crc kubenswrapper[4718]: I1206 04:23:53.940397 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq5w4\" (UniqueName: \"kubernetes.io/projected/147bc60c-8979-462d-bfc5-8460b289738d-kube-api-access-vq5w4\") pod \"keystone-operator-index-hql9d\" (UID: \"147bc60c-8979-462d-bfc5-8460b289738d\") " pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:23:53 crc kubenswrapper[4718]: I1206 04:23:53.960282 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq5w4\" (UniqueName: \"kubernetes.io/projected/147bc60c-8979-462d-bfc5-8460b289738d-kube-api-access-vq5w4\") pod \"keystone-operator-index-hql9d\" (UID: \"147bc60c-8979-462d-bfc5-8460b289738d\") " pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:23:54 crc kubenswrapper[4718]: I1206 04:23:54.003526 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:23:54 crc kubenswrapper[4718]: I1206 04:23:54.212119 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-hql9d"] Dec 06 04:23:55 crc kubenswrapper[4718]: I1206 04:23:55.210793 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hql9d" event={"ID":"147bc60c-8979-462d-bfc5-8460b289738d","Type":"ContainerStarted","Data":"b232968272c927c011084e0038f6ac71a0bedb88a94078a8c726e017065e5c98"} Dec 06 04:23:57 crc kubenswrapper[4718]: I1206 04:23:57.877204 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:23:57 crc kubenswrapper[4718]: I1206 04:23:57.878489 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:24:00 crc kubenswrapper[4718]: I1206 04:24:00.249334 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hql9d" event={"ID":"147bc60c-8979-462d-bfc5-8460b289738d","Type":"ContainerStarted","Data":"d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4"} Dec 06 04:24:00 crc kubenswrapper[4718]: I1206 04:24:00.267894 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-hql9d" podStartSLOduration=2.307836947 podStartE2EDuration="7.267871272s" podCreationTimestamp="2025-12-06 04:23:53 +0000 UTC" firstStartedPulling="2025-12-06 04:23:54.231522938 +0000 UTC m=+1023.237228099" lastFinishedPulling="2025-12-06 04:23:59.191557253 +0000 UTC m=+1028.197262424" observedRunningTime="2025-12-06 04:24:00.267124737 +0000 UTC m=+1029.272829938" watchObservedRunningTime="2025-12-06 04:24:00.267871272 +0000 UTC m=+1029.273576453" Dec 06 04:24:01 crc kubenswrapper[4718]: I1206 04:24:01.260078 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"e9868af3-c211-4dde-9701-669e579f484b","Type":"ContainerStarted","Data":"1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc"} Dec 06 04:24:04 crc kubenswrapper[4718]: I1206 04:24:04.003714 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:24:04 crc kubenswrapper[4718]: I1206 04:24:04.004182 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:24:04 crc kubenswrapper[4718]: I1206 04:24:04.043351 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:24:04 crc kubenswrapper[4718]: I1206 04:24:04.320399 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.488623 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh"] Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.490362 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.492597 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-69frt" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.535850 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh"] Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.649304 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtbc5\" (UniqueName: \"kubernetes.io/projected/092d522a-d186-4cfc-8f7e-1a337b4463ba-kube-api-access-rtbc5\") pod \"8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.649391 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-bundle\") pod \"8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.649461 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-util\") pod \"8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.751169 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-bundle\") pod \"8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.751369 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-util\") pod \"8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.751466 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtbc5\" (UniqueName: \"kubernetes.io/projected/092d522a-d186-4cfc-8f7e-1a337b4463ba-kube-api-access-rtbc5\") pod \"8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.752163 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-bundle\") pod \"8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.752335 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-util\") pod \"8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.775535 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtbc5\" (UniqueName: \"kubernetes.io/projected/092d522a-d186-4cfc-8f7e-1a337b4463ba-kube-api-access-rtbc5\") pod \"8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:06 crc kubenswrapper[4718]: I1206 04:24:06.826881 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:07 crc kubenswrapper[4718]: I1206 04:24:07.062869 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh"] Dec 06 04:24:07 crc kubenswrapper[4718]: I1206 04:24:07.307134 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" event={"ID":"092d522a-d186-4cfc-8f7e-1a337b4463ba","Type":"ContainerStarted","Data":"5de89b4026a57af3c1e76a6b8ef008b05b02e357800f04d707cf96ee442ce5e4"} Dec 06 04:24:07 crc kubenswrapper[4718]: I1206 04:24:07.307184 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" event={"ID":"092d522a-d186-4cfc-8f7e-1a337b4463ba","Type":"ContainerStarted","Data":"6759c443bcd8bf2cb9a8f65387035b78fb5db201085fe62129fcebd12b8db2b4"} Dec 06 04:24:08 crc kubenswrapper[4718]: I1206 04:24:08.314055 4718 generic.go:334] "Generic (PLEG): container finished" podID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerID="5de89b4026a57af3c1e76a6b8ef008b05b02e357800f04d707cf96ee442ce5e4" exitCode=0 Dec 06 04:24:08 crc kubenswrapper[4718]: I1206 04:24:08.314108 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" event={"ID":"092d522a-d186-4cfc-8f7e-1a337b4463ba","Type":"ContainerDied","Data":"5de89b4026a57af3c1e76a6b8ef008b05b02e357800f04d707cf96ee442ce5e4"} Dec 06 04:24:09 crc kubenswrapper[4718]: I1206 04:24:09.323073 4718 generic.go:334] "Generic (PLEG): container finished" podID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerID="82e89a00ff8b153a78b8b13c5f8c6095db280876aa36864f4caa0d42af26654a" exitCode=0 Dec 06 04:24:09 crc kubenswrapper[4718]: I1206 04:24:09.323114 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" event={"ID":"092d522a-d186-4cfc-8f7e-1a337b4463ba","Type":"ContainerDied","Data":"82e89a00ff8b153a78b8b13c5f8c6095db280876aa36864f4caa0d42af26654a"} Dec 06 04:24:10 crc kubenswrapper[4718]: I1206 04:24:10.330892 4718 generic.go:334] "Generic (PLEG): container finished" podID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerID="715de72658276c0e32f61d47b2bec2d5a432ab5fd4fd2816bfe6e6b17d4ba802" exitCode=0 Dec 06 04:24:10 crc kubenswrapper[4718]: I1206 04:24:10.330933 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" event={"ID":"092d522a-d186-4cfc-8f7e-1a337b4463ba","Type":"ContainerDied","Data":"715de72658276c0e32f61d47b2bec2d5a432ab5fd4fd2816bfe6e6b17d4ba802"} Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.710272 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.825569 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-util\") pod \"092d522a-d186-4cfc-8f7e-1a337b4463ba\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.826046 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-bundle\") pod \"092d522a-d186-4cfc-8f7e-1a337b4463ba\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.826080 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtbc5\" (UniqueName: \"kubernetes.io/projected/092d522a-d186-4cfc-8f7e-1a337b4463ba-kube-api-access-rtbc5\") pod \"092d522a-d186-4cfc-8f7e-1a337b4463ba\" (UID: \"092d522a-d186-4cfc-8f7e-1a337b4463ba\") " Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.826964 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-bundle" (OuterVolumeSpecName: "bundle") pod "092d522a-d186-4cfc-8f7e-1a337b4463ba" (UID: "092d522a-d186-4cfc-8f7e-1a337b4463ba"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.834198 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/092d522a-d186-4cfc-8f7e-1a337b4463ba-kube-api-access-rtbc5" (OuterVolumeSpecName: "kube-api-access-rtbc5") pod "092d522a-d186-4cfc-8f7e-1a337b4463ba" (UID: "092d522a-d186-4cfc-8f7e-1a337b4463ba"). InnerVolumeSpecName "kube-api-access-rtbc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.839919 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-util" (OuterVolumeSpecName: "util") pod "092d522a-d186-4cfc-8f7e-1a337b4463ba" (UID: "092d522a-d186-4cfc-8f7e-1a337b4463ba"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.927994 4718 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.928056 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtbc5\" (UniqueName: \"kubernetes.io/projected/092d522a-d186-4cfc-8f7e-1a337b4463ba-kube-api-access-rtbc5\") on node \"crc\" DevicePath \"\"" Dec 06 04:24:11 crc kubenswrapper[4718]: I1206 04:24:11.928078 4718 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/092d522a-d186-4cfc-8f7e-1a337b4463ba-util\") on node \"crc\" DevicePath \"\"" Dec 06 04:24:12 crc kubenswrapper[4718]: I1206 04:24:12.377672 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" event={"ID":"092d522a-d186-4cfc-8f7e-1a337b4463ba","Type":"ContainerDied","Data":"6759c443bcd8bf2cb9a8f65387035b78fb5db201085fe62129fcebd12b8db2b4"} Dec 06 04:24:12 crc kubenswrapper[4718]: I1206 04:24:12.377712 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6759c443bcd8bf2cb9a8f65387035b78fb5db201085fe62129fcebd12b8db2b4" Dec 06 04:24:12 crc kubenswrapper[4718]: I1206 04:24:12.377787 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.308039 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd"] Dec 06 04:24:26 crc kubenswrapper[4718]: E1206 04:24:26.308831 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerName="util" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.308864 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerName="util" Dec 06 04:24:26 crc kubenswrapper[4718]: E1206 04:24:26.308875 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerName="extract" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.308881 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerName="extract" Dec 06 04:24:26 crc kubenswrapper[4718]: E1206 04:24:26.308894 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerName="pull" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.308900 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerName="pull" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.309019 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="092d522a-d186-4cfc-8f7e-1a337b4463ba" containerName="extract" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.309464 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.312293 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.312497 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-j8pkt" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.319148 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd"] Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.451601 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-webhook-cert\") pod \"keystone-operator-controller-manager-94f9cff5f-65dbd\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.451675 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqp9z\" (UniqueName: \"kubernetes.io/projected/8277a3e3-eae1-486b-936d-e921d8991932-kube-api-access-bqp9z\") pod \"keystone-operator-controller-manager-94f9cff5f-65dbd\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.451724 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-apiservice-cert\") pod \"keystone-operator-controller-manager-94f9cff5f-65dbd\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.553778 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-webhook-cert\") pod \"keystone-operator-controller-manager-94f9cff5f-65dbd\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.553919 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqp9z\" (UniqueName: \"kubernetes.io/projected/8277a3e3-eae1-486b-936d-e921d8991932-kube-api-access-bqp9z\") pod \"keystone-operator-controller-manager-94f9cff5f-65dbd\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.554031 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-apiservice-cert\") pod \"keystone-operator-controller-manager-94f9cff5f-65dbd\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.560137 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-apiservice-cert\") pod \"keystone-operator-controller-manager-94f9cff5f-65dbd\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.560894 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-webhook-cert\") pod \"keystone-operator-controller-manager-94f9cff5f-65dbd\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.571833 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqp9z\" (UniqueName: \"kubernetes.io/projected/8277a3e3-eae1-486b-936d-e921d8991932-kube-api-access-bqp9z\") pod \"keystone-operator-controller-manager-94f9cff5f-65dbd\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:26 crc kubenswrapper[4718]: I1206 04:24:26.626774 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:27 crc kubenswrapper[4718]: I1206 04:24:27.057837 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd"] Dec 06 04:24:27 crc kubenswrapper[4718]: I1206 04:24:27.479720 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" event={"ID":"8277a3e3-eae1-486b-936d-e921d8991932","Type":"ContainerStarted","Data":"9e2414b1993cb4a69e32586595c37081528356472d3280e20571b8734c9eb71d"} Dec 06 04:24:27 crc kubenswrapper[4718]: I1206 04:24:27.877666 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:24:27 crc kubenswrapper[4718]: I1206 04:24:27.878060 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:24:27 crc kubenswrapper[4718]: I1206 04:24:27.878382 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:24:27 crc kubenswrapper[4718]: I1206 04:24:27.879555 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5492a0b85bf17b4550c5ff3ce07a61828c2b6f5eb097631d3cdfc69dc9a69c85"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:24:27 crc kubenswrapper[4718]: I1206 04:24:27.879820 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://5492a0b85bf17b4550c5ff3ce07a61828c2b6f5eb097631d3cdfc69dc9a69c85" gracePeriod=600 Dec 06 04:24:28 crc kubenswrapper[4718]: I1206 04:24:28.488223 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="5492a0b85bf17b4550c5ff3ce07a61828c2b6f5eb097631d3cdfc69dc9a69c85" exitCode=0 Dec 06 04:24:28 crc kubenswrapper[4718]: I1206 04:24:28.488586 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"5492a0b85bf17b4550c5ff3ce07a61828c2b6f5eb097631d3cdfc69dc9a69c85"} Dec 06 04:24:28 crc kubenswrapper[4718]: I1206 04:24:28.488650 4718 scope.go:117] "RemoveContainer" containerID="5d80f8884240523b385c679ab95f41f5a9a967b7f7ed7e01c46ae96cdb397ec1" Dec 06 04:24:29 crc kubenswrapper[4718]: I1206 04:24:29.497263 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"8d135c41c1c0b733890b2561029cffaa9956d80b356ad7dc0217957931de3b87"} Dec 06 04:24:31 crc kubenswrapper[4718]: I1206 04:24:31.517994 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" event={"ID":"8277a3e3-eae1-486b-936d-e921d8991932","Type":"ContainerStarted","Data":"6e61a0ec9d7b0d5ca50f9ae775dd6feea92c57bdd50ee7297f67732d5544e518"} Dec 06 04:24:31 crc kubenswrapper[4718]: I1206 04:24:31.518717 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:31 crc kubenswrapper[4718]: I1206 04:24:31.537052 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" podStartSLOduration=1.7514346 podStartE2EDuration="5.537036461s" podCreationTimestamp="2025-12-06 04:24:26 +0000 UTC" firstStartedPulling="2025-12-06 04:24:27.07570095 +0000 UTC m=+1056.081406121" lastFinishedPulling="2025-12-06 04:24:30.861302821 +0000 UTC m=+1059.867007982" observedRunningTime="2025-12-06 04:24:31.536278944 +0000 UTC m=+1060.541984125" watchObservedRunningTime="2025-12-06 04:24:31.537036461 +0000 UTC m=+1060.542741622" Dec 06 04:24:33 crc kubenswrapper[4718]: I1206 04:24:33.538965 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9868af3-c211-4dde-9701-669e579f484b" containerID="1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc" exitCode=0 Dec 06 04:24:33 crc kubenswrapper[4718]: I1206 04:24:33.539033 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"e9868af3-c211-4dde-9701-669e579f484b","Type":"ContainerDied","Data":"1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc"} Dec 06 04:24:34 crc kubenswrapper[4718]: I1206 04:24:34.547419 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"e9868af3-c211-4dde-9701-669e579f484b","Type":"ContainerStarted","Data":"c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096"} Dec 06 04:24:34 crc kubenswrapper[4718]: I1206 04:24:34.547873 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:24:34 crc kubenswrapper[4718]: I1206 04:24:34.579785 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.532569592 podStartE2EDuration="44.579744641s" podCreationTimestamp="2025-12-06 04:23:50 +0000 UTC" firstStartedPulling="2025-12-06 04:23:52.712526137 +0000 UTC m=+1021.718231308" lastFinishedPulling="2025-12-06 04:23:59.759701196 +0000 UTC m=+1028.765406357" observedRunningTime="2025-12-06 04:24:34.573795909 +0000 UTC m=+1063.579501150" watchObservedRunningTime="2025-12-06 04:24:34.579744641 +0000 UTC m=+1063.585449802" Dec 06 04:24:36 crc kubenswrapper[4718]: I1206 04:24:36.633135 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.542671 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6"] Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.545268 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.550318 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.555859 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6"] Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.584711 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt57d\" (UniqueName: \"kubernetes.io/projected/65d645a2-187b-4fb8-af34-5d1bad8d580b-kube-api-access-gt57d\") pod \"keystone-385d-account-create-update-cgvr6\" (UID: \"65d645a2-187b-4fb8-af34-5d1bad8d580b\") " pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.584758 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d645a2-187b-4fb8-af34-5d1bad8d580b-operator-scripts\") pod \"keystone-385d-account-create-update-cgvr6\" (UID: \"65d645a2-187b-4fb8-af34-5d1bad8d580b\") " pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.627545 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-5jl2m"] Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.628346 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.636287 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-5jl2m"] Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.686052 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt57d\" (UniqueName: \"kubernetes.io/projected/65d645a2-187b-4fb8-af34-5d1bad8d580b-kube-api-access-gt57d\") pod \"keystone-385d-account-create-update-cgvr6\" (UID: \"65d645a2-187b-4fb8-af34-5d1bad8d580b\") " pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.686113 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d645a2-187b-4fb8-af34-5d1bad8d580b-operator-scripts\") pod \"keystone-385d-account-create-update-cgvr6\" (UID: \"65d645a2-187b-4fb8-af34-5d1bad8d580b\") " pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.687205 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d645a2-187b-4fb8-af34-5d1bad8d580b-operator-scripts\") pod \"keystone-385d-account-create-update-cgvr6\" (UID: \"65d645a2-187b-4fb8-af34-5d1bad8d580b\") " pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.703980 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt57d\" (UniqueName: \"kubernetes.io/projected/65d645a2-187b-4fb8-af34-5d1bad8d580b-kube-api-access-gt57d\") pod \"keystone-385d-account-create-update-cgvr6\" (UID: \"65d645a2-187b-4fb8-af34-5d1bad8d580b\") " pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.787711 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vrzr\" (UniqueName: \"kubernetes.io/projected/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-kube-api-access-6vrzr\") pod \"keystone-db-create-5jl2m\" (UID: \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\") " pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.787759 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-operator-scripts\") pod \"keystone-db-create-5jl2m\" (UID: \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\") " pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.867853 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.888891 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vrzr\" (UniqueName: \"kubernetes.io/projected/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-kube-api-access-6vrzr\") pod \"keystone-db-create-5jl2m\" (UID: \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\") " pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.888961 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-operator-scripts\") pod \"keystone-db-create-5jl2m\" (UID: \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\") " pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.890012 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-operator-scripts\") pod \"keystone-db-create-5jl2m\" (UID: \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\") " pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.922101 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vrzr\" (UniqueName: \"kubernetes.io/projected/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-kube-api-access-6vrzr\") pod \"keystone-db-create-5jl2m\" (UID: \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\") " pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:45 crc kubenswrapper[4718]: I1206 04:24:45.940823 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:46 crc kubenswrapper[4718]: I1206 04:24:46.355023 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6"] Dec 06 04:24:46 crc kubenswrapper[4718]: W1206 04:24:46.405053 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4430b20_c353_47e4_9fb3_ea1aaeb35a16.slice/crio-a25ea6ffccb62a0fe9e3a1d1347399a4d77141e7e3d14082cc33f30208d4f85f WatchSource:0}: Error finding container a25ea6ffccb62a0fe9e3a1d1347399a4d77141e7e3d14082cc33f30208d4f85f: Status 404 returned error can't find the container with id a25ea6ffccb62a0fe9e3a1d1347399a4d77141e7e3d14082cc33f30208d4f85f Dec 06 04:24:46 crc kubenswrapper[4718]: I1206 04:24:46.407293 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-5jl2m"] Dec 06 04:24:46 crc kubenswrapper[4718]: I1206 04:24:46.640880 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" event={"ID":"65d645a2-187b-4fb8-af34-5d1bad8d580b","Type":"ContainerStarted","Data":"d4097c8b997fb1210a6abf71344ec95d75701d2c96a10b2355a22ac2cb4cfb58"} Dec 06 04:24:46 crc kubenswrapper[4718]: I1206 04:24:46.642657 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" event={"ID":"65d645a2-187b-4fb8-af34-5d1bad8d580b","Type":"ContainerStarted","Data":"c548ba09507e768f43b8401192608a17884a605c3d86c969b3d965fa56ce01c0"} Dec 06 04:24:46 crc kubenswrapper[4718]: I1206 04:24:46.642745 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-5jl2m" event={"ID":"e4430b20-c353-47e4-9fb3-ea1aaeb35a16","Type":"ContainerStarted","Data":"c4e319dd489da8e7cb93d75fc7df1f3be46c078698a0dc8e8c695fe1c20bcd42"} Dec 06 04:24:46 crc kubenswrapper[4718]: I1206 04:24:46.642774 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-5jl2m" event={"ID":"e4430b20-c353-47e4-9fb3-ea1aaeb35a16","Type":"ContainerStarted","Data":"a25ea6ffccb62a0fe9e3a1d1347399a4d77141e7e3d14082cc33f30208d4f85f"} Dec 06 04:24:46 crc kubenswrapper[4718]: I1206 04:24:46.658152 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" podStartSLOduration=1.658132083 podStartE2EDuration="1.658132083s" podCreationTimestamp="2025-12-06 04:24:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:24:46.654108487 +0000 UTC m=+1075.659813668" watchObservedRunningTime="2025-12-06 04:24:46.658132083 +0000 UTC m=+1075.663837244" Dec 06 04:24:46 crc kubenswrapper[4718]: I1206 04:24:46.675518 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-db-create-5jl2m" podStartSLOduration=1.675488413 podStartE2EDuration="1.675488413s" podCreationTimestamp="2025-12-06 04:24:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:24:46.668554157 +0000 UTC m=+1075.674259328" watchObservedRunningTime="2025-12-06 04:24:46.675488413 +0000 UTC m=+1075.681193614" Dec 06 04:24:47 crc kubenswrapper[4718]: I1206 04:24:47.658459 4718 generic.go:334] "Generic (PLEG): container finished" podID="e4430b20-c353-47e4-9fb3-ea1aaeb35a16" containerID="c4e319dd489da8e7cb93d75fc7df1f3be46c078698a0dc8e8c695fe1c20bcd42" exitCode=0 Dec 06 04:24:47 crc kubenswrapper[4718]: I1206 04:24:47.658902 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-5jl2m" event={"ID":"e4430b20-c353-47e4-9fb3-ea1aaeb35a16","Type":"ContainerDied","Data":"c4e319dd489da8e7cb93d75fc7df1f3be46c078698a0dc8e8c695fe1c20bcd42"} Dec 06 04:24:47 crc kubenswrapper[4718]: I1206 04:24:47.660473 4718 generic.go:334] "Generic (PLEG): container finished" podID="65d645a2-187b-4fb8-af34-5d1bad8d580b" containerID="d4097c8b997fb1210a6abf71344ec95d75701d2c96a10b2355a22ac2cb4cfb58" exitCode=0 Dec 06 04:24:47 crc kubenswrapper[4718]: I1206 04:24:47.660512 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" event={"ID":"65d645a2-187b-4fb8-af34-5d1bad8d580b","Type":"ContainerDied","Data":"d4097c8b997fb1210a6abf71344ec95d75701d2c96a10b2355a22ac2cb4cfb58"} Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.208884 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.218072 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.336668 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt57d\" (UniqueName: \"kubernetes.io/projected/65d645a2-187b-4fb8-af34-5d1bad8d580b-kube-api-access-gt57d\") pod \"65d645a2-187b-4fb8-af34-5d1bad8d580b\" (UID: \"65d645a2-187b-4fb8-af34-5d1bad8d580b\") " Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.336900 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-operator-scripts\") pod \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\" (UID: \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\") " Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.336949 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vrzr\" (UniqueName: \"kubernetes.io/projected/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-kube-api-access-6vrzr\") pod \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\" (UID: \"e4430b20-c353-47e4-9fb3-ea1aaeb35a16\") " Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.337012 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d645a2-187b-4fb8-af34-5d1bad8d580b-operator-scripts\") pod \"65d645a2-187b-4fb8-af34-5d1bad8d580b\" (UID: \"65d645a2-187b-4fb8-af34-5d1bad8d580b\") " Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.337637 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e4430b20-c353-47e4-9fb3-ea1aaeb35a16" (UID: "e4430b20-c353-47e4-9fb3-ea1aaeb35a16"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.337691 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65d645a2-187b-4fb8-af34-5d1bad8d580b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "65d645a2-187b-4fb8-af34-5d1bad8d580b" (UID: "65d645a2-187b-4fb8-af34-5d1bad8d580b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.349753 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65d645a2-187b-4fb8-af34-5d1bad8d580b-kube-api-access-gt57d" (OuterVolumeSpecName: "kube-api-access-gt57d") pod "65d645a2-187b-4fb8-af34-5d1bad8d580b" (UID: "65d645a2-187b-4fb8-af34-5d1bad8d580b"). InnerVolumeSpecName "kube-api-access-gt57d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.351381 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-kube-api-access-6vrzr" (OuterVolumeSpecName: "kube-api-access-6vrzr") pod "e4430b20-c353-47e4-9fb3-ea1aaeb35a16" (UID: "e4430b20-c353-47e4-9fb3-ea1aaeb35a16"). InnerVolumeSpecName "kube-api-access-6vrzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.438696 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt57d\" (UniqueName: \"kubernetes.io/projected/65d645a2-187b-4fb8-af34-5d1bad8d580b-kube-api-access-gt57d\") on node \"crc\" DevicePath \"\"" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.438736 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.438750 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vrzr\" (UniqueName: \"kubernetes.io/projected/e4430b20-c353-47e4-9fb3-ea1aaeb35a16-kube-api-access-6vrzr\") on node \"crc\" DevicePath \"\"" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.438760 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d645a2-187b-4fb8-af34-5d1bad8d580b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.691378 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-5jl2m" event={"ID":"e4430b20-c353-47e4-9fb3-ea1aaeb35a16","Type":"ContainerDied","Data":"a25ea6ffccb62a0fe9e3a1d1347399a4d77141e7e3d14082cc33f30208d4f85f"} Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.691442 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a25ea6ffccb62a0fe9e3a1d1347399a4d77141e7e3d14082cc33f30208d4f85f" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.691611 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-5jl2m" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.695346 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" event={"ID":"65d645a2-187b-4fb8-af34-5d1bad8d580b","Type":"ContainerDied","Data":"c548ba09507e768f43b8401192608a17884a605c3d86c969b3d965fa56ce01c0"} Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.695377 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c548ba09507e768f43b8401192608a17884a605c3d86c969b3d965fa56ce01c0" Dec 06 04:24:49 crc kubenswrapper[4718]: I1206 04:24:49.695457 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.220483 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.775052 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-qkxrs"] Dec 06 04:24:52 crc kubenswrapper[4718]: E1206 04:24:52.775505 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4430b20-c353-47e4-9fb3-ea1aaeb35a16" containerName="mariadb-database-create" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.775516 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4430b20-c353-47e4-9fb3-ea1aaeb35a16" containerName="mariadb-database-create" Dec 06 04:24:52 crc kubenswrapper[4718]: E1206 04:24:52.775530 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d645a2-187b-4fb8-af34-5d1bad8d580b" containerName="mariadb-account-create-update" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.775536 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d645a2-187b-4fb8-af34-5d1bad8d580b" containerName="mariadb-account-create-update" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.775640 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d645a2-187b-4fb8-af34-5d1bad8d580b" containerName="mariadb-account-create-update" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.775654 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4430b20-c353-47e4-9fb3-ea1aaeb35a16" containerName="mariadb-database-create" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.776054 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.782330 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-86krp" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.782366 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.782689 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.783932 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-qkxrs"] Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.784205 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.888937 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cr54\" (UniqueName: \"kubernetes.io/projected/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-kube-api-access-7cr54\") pod \"keystone-db-sync-qkxrs\" (UID: \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\") " pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.889271 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-config-data\") pod \"keystone-db-sync-qkxrs\" (UID: \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\") " pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.991137 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-config-data\") pod \"keystone-db-sync-qkxrs\" (UID: \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\") " pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:24:52 crc kubenswrapper[4718]: I1206 04:24:52.991478 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cr54\" (UniqueName: \"kubernetes.io/projected/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-kube-api-access-7cr54\") pod \"keystone-db-sync-qkxrs\" (UID: \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\") " pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:24:53 crc kubenswrapper[4718]: I1206 04:24:53.001772 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-config-data\") pod \"keystone-db-sync-qkxrs\" (UID: \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\") " pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:24:53 crc kubenswrapper[4718]: I1206 04:24:53.018790 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cr54\" (UniqueName: \"kubernetes.io/projected/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-kube-api-access-7cr54\") pod \"keystone-db-sync-qkxrs\" (UID: \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\") " pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:24:53 crc kubenswrapper[4718]: I1206 04:24:53.092111 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:24:53 crc kubenswrapper[4718]: I1206 04:24:53.530077 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-qkxrs"] Dec 06 04:24:53 crc kubenswrapper[4718]: W1206 04:24:53.535821 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12732d17_ee22_4bef_ab33_4ef7ebd1b6f9.slice/crio-273cff7b23b2933af6e89bcfe406a18afd4d22db56be26aa95e476212c54245a WatchSource:0}: Error finding container 273cff7b23b2933af6e89bcfe406a18afd4d22db56be26aa95e476212c54245a: Status 404 returned error can't find the container with id 273cff7b23b2933af6e89bcfe406a18afd4d22db56be26aa95e476212c54245a Dec 06 04:24:53 crc kubenswrapper[4718]: I1206 04:24:53.725862 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" event={"ID":"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9","Type":"ContainerStarted","Data":"273cff7b23b2933af6e89bcfe406a18afd4d22db56be26aa95e476212c54245a"} Dec 06 04:25:00 crc kubenswrapper[4718]: I1206 04:25:00.777446 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" event={"ID":"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9","Type":"ContainerStarted","Data":"ba3dc75586f7a5125e1d161f027763267ee39b7cbd3255acad43fa1b1aa5dd93"} Dec 06 04:25:00 crc kubenswrapper[4718]: I1206 04:25:00.805773 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" podStartSLOduration=1.869574923 podStartE2EDuration="8.80574479s" podCreationTimestamp="2025-12-06 04:24:52 +0000 UTC" firstStartedPulling="2025-12-06 04:24:53.538049709 +0000 UTC m=+1082.543754870" lastFinishedPulling="2025-12-06 04:25:00.474219576 +0000 UTC m=+1089.479924737" observedRunningTime="2025-12-06 04:25:00.798291767 +0000 UTC m=+1089.803996928" watchObservedRunningTime="2025-12-06 04:25:00.80574479 +0000 UTC m=+1089.811449981" Dec 06 04:25:04 crc kubenswrapper[4718]: I1206 04:25:04.808497 4718 generic.go:334] "Generic (PLEG): container finished" podID="12732d17-ee22-4bef-ab33-4ef7ebd1b6f9" containerID="ba3dc75586f7a5125e1d161f027763267ee39b7cbd3255acad43fa1b1aa5dd93" exitCode=0 Dec 06 04:25:04 crc kubenswrapper[4718]: I1206 04:25:04.808564 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" event={"ID":"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9","Type":"ContainerDied","Data":"ba3dc75586f7a5125e1d161f027763267ee39b7cbd3255acad43fa1b1aa5dd93"} Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.109665 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.184994 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cr54\" (UniqueName: \"kubernetes.io/projected/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-kube-api-access-7cr54\") pod \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\" (UID: \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\") " Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.185076 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-config-data\") pod \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\" (UID: \"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9\") " Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.195548 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-kube-api-access-7cr54" (OuterVolumeSpecName: "kube-api-access-7cr54") pod "12732d17-ee22-4bef-ab33-4ef7ebd1b6f9" (UID: "12732d17-ee22-4bef-ab33-4ef7ebd1b6f9"). InnerVolumeSpecName "kube-api-access-7cr54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.233812 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-config-data" (OuterVolumeSpecName: "config-data") pod "12732d17-ee22-4bef-ab33-4ef7ebd1b6f9" (UID: "12732d17-ee22-4bef-ab33-4ef7ebd1b6f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.286724 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cr54\" (UniqueName: \"kubernetes.io/projected/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-kube-api-access-7cr54\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.286768 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.827073 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" event={"ID":"12732d17-ee22-4bef-ab33-4ef7ebd1b6f9","Type":"ContainerDied","Data":"273cff7b23b2933af6e89bcfe406a18afd4d22db56be26aa95e476212c54245a"} Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.827104 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-qkxrs" Dec 06 04:25:06 crc kubenswrapper[4718]: I1206 04:25:06.827111 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="273cff7b23b2933af6e89bcfe406a18afd4d22db56be26aa95e476212c54245a" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.020580 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-vcmmj"] Dec 06 04:25:07 crc kubenswrapper[4718]: E1206 04:25:07.020850 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12732d17-ee22-4bef-ab33-4ef7ebd1b6f9" containerName="keystone-db-sync" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.020866 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="12732d17-ee22-4bef-ab33-4ef7ebd1b6f9" containerName="keystone-db-sync" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.020967 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="12732d17-ee22-4bef-ab33-4ef7ebd1b6f9" containerName="keystone-db-sync" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.021419 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.023419 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.023949 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.024134 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-vcmmj"] Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.024165 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.027189 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.027325 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-86krp" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.096222 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv6sb\" (UniqueName: \"kubernetes.io/projected/db3759d7-a87d-492b-a711-72550fc419bd-kube-api-access-dv6sb\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.096354 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-config-data\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.096391 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-fernet-keys\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.096417 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-scripts\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.096440 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-credential-keys\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.197573 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-config-data\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.198492 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-fernet-keys\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.198930 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-scripts\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.198973 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-credential-keys\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.199175 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv6sb\" (UniqueName: \"kubernetes.io/projected/db3759d7-a87d-492b-a711-72550fc419bd-kube-api-access-dv6sb\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.202290 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-config-data\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.202688 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-fernet-keys\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.203133 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-credential-keys\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.203403 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-scripts\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.220339 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv6sb\" (UniqueName: \"kubernetes.io/projected/db3759d7-a87d-492b-a711-72550fc419bd-kube-api-access-dv6sb\") pod \"keystone-bootstrap-vcmmj\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.346957 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.579972 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-vcmmj"] Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.851387 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" event={"ID":"db3759d7-a87d-492b-a711-72550fc419bd","Type":"ContainerStarted","Data":"a1da07bd695de999dc679e8aea7d091aa409047b1878a5b0980ba4be4bc848f7"} Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.851449 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" event={"ID":"db3759d7-a87d-492b-a711-72550fc419bd","Type":"ContainerStarted","Data":"fdfdd43954ce9b0446f2eb8c8c4f40b4861267dba71cf9aedd78dd3741c7fa55"} Dec 06 04:25:07 crc kubenswrapper[4718]: I1206 04:25:07.872138 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" podStartSLOduration=1.872118451 podStartE2EDuration="1.872118451s" podCreationTimestamp="2025-12-06 04:25:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:25:07.868328322 +0000 UTC m=+1096.874033493" watchObservedRunningTime="2025-12-06 04:25:07.872118451 +0000 UTC m=+1096.877823622" Dec 06 04:25:11 crc kubenswrapper[4718]: I1206 04:25:11.883787 4718 generic.go:334] "Generic (PLEG): container finished" podID="db3759d7-a87d-492b-a711-72550fc419bd" containerID="a1da07bd695de999dc679e8aea7d091aa409047b1878a5b0980ba4be4bc848f7" exitCode=0 Dec 06 04:25:11 crc kubenswrapper[4718]: I1206 04:25:11.883842 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" event={"ID":"db3759d7-a87d-492b-a711-72550fc419bd","Type":"ContainerDied","Data":"a1da07bd695de999dc679e8aea7d091aa409047b1878a5b0980ba4be4bc848f7"} Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.248211 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.440402 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv6sb\" (UniqueName: \"kubernetes.io/projected/db3759d7-a87d-492b-a711-72550fc419bd-kube-api-access-dv6sb\") pod \"db3759d7-a87d-492b-a711-72550fc419bd\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.440566 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-credential-keys\") pod \"db3759d7-a87d-492b-a711-72550fc419bd\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.440620 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-scripts\") pod \"db3759d7-a87d-492b-a711-72550fc419bd\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.440652 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-fernet-keys\") pod \"db3759d7-a87d-492b-a711-72550fc419bd\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.440701 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-config-data\") pod \"db3759d7-a87d-492b-a711-72550fc419bd\" (UID: \"db3759d7-a87d-492b-a711-72550fc419bd\") " Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.445749 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "db3759d7-a87d-492b-a711-72550fc419bd" (UID: "db3759d7-a87d-492b-a711-72550fc419bd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.446205 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-scripts" (OuterVolumeSpecName: "scripts") pod "db3759d7-a87d-492b-a711-72550fc419bd" (UID: "db3759d7-a87d-492b-a711-72550fc419bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.446394 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db3759d7-a87d-492b-a711-72550fc419bd-kube-api-access-dv6sb" (OuterVolumeSpecName: "kube-api-access-dv6sb") pod "db3759d7-a87d-492b-a711-72550fc419bd" (UID: "db3759d7-a87d-492b-a711-72550fc419bd"). InnerVolumeSpecName "kube-api-access-dv6sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.454019 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "db3759d7-a87d-492b-a711-72550fc419bd" (UID: "db3759d7-a87d-492b-a711-72550fc419bd"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.474379 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-config-data" (OuterVolumeSpecName: "config-data") pod "db3759d7-a87d-492b-a711-72550fc419bd" (UID: "db3759d7-a87d-492b-a711-72550fc419bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.542640 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.542682 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.542700 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.542717 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3759d7-a87d-492b-a711-72550fc419bd-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.542735 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv6sb\" (UniqueName: \"kubernetes.io/projected/db3759d7-a87d-492b-a711-72550fc419bd-kube-api-access-dv6sb\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.903687 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" event={"ID":"db3759d7-a87d-492b-a711-72550fc419bd","Type":"ContainerDied","Data":"fdfdd43954ce9b0446f2eb8c8c4f40b4861267dba71cf9aedd78dd3741c7fa55"} Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.903997 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdfdd43954ce9b0446f2eb8c8c4f40b4861267dba71cf9aedd78dd3741c7fa55" Dec 06 04:25:13 crc kubenswrapper[4718]: I1206 04:25:13.903784 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-vcmmj" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.094443 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq"] Dec 06 04:25:14 crc kubenswrapper[4718]: E1206 04:25:14.095064 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db3759d7-a87d-492b-a711-72550fc419bd" containerName="keystone-bootstrap" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.095120 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="db3759d7-a87d-492b-a711-72550fc419bd" containerName="keystone-bootstrap" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.095304 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="db3759d7-a87d-492b-a711-72550fc419bd" containerName="keystone-bootstrap" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.095783 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.097988 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.098056 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-86krp" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.098352 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.100423 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.113998 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq"] Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.252594 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-credential-keys\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.252635 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsbch\" (UniqueName: \"kubernetes.io/projected/b4c57765-c010-4794-8e32-11340a76ea67-kube-api-access-nsbch\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.252685 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-scripts\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.252710 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-config-data\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.252902 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-fernet-keys\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.354912 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-fernet-keys\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.355120 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-credential-keys\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.355173 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsbch\" (UniqueName: \"kubernetes.io/projected/b4c57765-c010-4794-8e32-11340a76ea67-kube-api-access-nsbch\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.355960 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-scripts\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.356698 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-config-data\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.360981 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-credential-keys\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.361379 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-fernet-keys\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.361619 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-scripts\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.362371 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-config-data\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.386862 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsbch\" (UniqueName: \"kubernetes.io/projected/b4c57765-c010-4794-8e32-11340a76ea67-kube-api-access-nsbch\") pod \"keystone-79fc8bbcc8-9wwdq\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.418644 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.865643 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq"] Dec 06 04:25:14 crc kubenswrapper[4718]: I1206 04:25:14.914727 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" event={"ID":"b4c57765-c010-4794-8e32-11340a76ea67","Type":"ContainerStarted","Data":"c93a9328532e55a7c200ff12728b028168edebd9b1bb416d225015be79440a34"} Dec 06 04:25:15 crc kubenswrapper[4718]: I1206 04:25:15.923974 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" event={"ID":"b4c57765-c010-4794-8e32-11340a76ea67","Type":"ContainerStarted","Data":"bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f"} Dec 06 04:25:15 crc kubenswrapper[4718]: I1206 04:25:15.924514 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:15 crc kubenswrapper[4718]: I1206 04:25:15.953405 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" podStartSLOduration=1.953390546 podStartE2EDuration="1.953390546s" podCreationTimestamp="2025-12-06 04:25:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:25:15.948159921 +0000 UTC m=+1104.953865082" watchObservedRunningTime="2025-12-06 04:25:15.953390546 +0000 UTC m=+1104.959095707" Dec 06 04:25:45 crc kubenswrapper[4718]: I1206 04:25:45.791599 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:46 crc kubenswrapper[4718]: E1206 04:25:46.601696 4718 log.go:32] "Failed when writing line to log file" err="http2: stream closed" path="/var/log/pods/keystone-kuttl-tests_keystone-79fc8bbcc8-9wwdq_b4c57765-c010-4794-8e32-11340a76ea67/keystone-api/0.log" line={} Dec 06 04:25:46 crc kubenswrapper[4718]: I1206 04:25:46.940256 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-c4fd8b875-2jbxs"] Dec 06 04:25:46 crc kubenswrapper[4718]: I1206 04:25:46.940962 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:46 crc kubenswrapper[4718]: I1206 04:25:46.958090 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-c4fd8b875-2jbxs"] Dec 06 04:25:47 crc kubenswrapper[4718]: E1206 04:25:47.024790 4718 log.go:32] "Failed when writing line to log file" err="http2: stream closed" path="/var/log/pods/keystone-kuttl-tests_keystone-79fc8bbcc8-9wwdq_b4c57765-c010-4794-8e32-11340a76ea67/keystone-api/0.log" line={} Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.085038 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.085096 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.085130 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.085298 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.085339 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lmjn\" (UniqueName: \"kubernetes.io/projected/6b518e06-ee13-41e3-8d3d-80139e9722a0-kube-api-access-5lmjn\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.186724 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.186788 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.186842 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.186943 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.186965 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lmjn\" (UniqueName: \"kubernetes.io/projected/6b518e06-ee13-41e3-8d3d-80139e9722a0-kube-api-access-5lmjn\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.192895 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.196008 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.201798 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.202625 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.214984 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lmjn\" (UniqueName: \"kubernetes.io/projected/6b518e06-ee13-41e3-8d3d-80139e9722a0-kube-api-access-5lmjn\") pod \"keystone-c4fd8b875-2jbxs\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.269840 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:47 crc kubenswrapper[4718]: I1206 04:25:47.785149 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-c4fd8b875-2jbxs"] Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.162530 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" event={"ID":"6b518e06-ee13-41e3-8d3d-80139e9722a0","Type":"ContainerStarted","Data":"439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea"} Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.162889 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" event={"ID":"6b518e06-ee13-41e3-8d3d-80139e9722a0","Type":"ContainerStarted","Data":"7c27d648c856b33fd956cf4877eda8d07d1cbad4c2b5ce7c5e5de8a6914a26f6"} Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.162978 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.200072 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" podStartSLOduration=2.200047239 podStartE2EDuration="2.200047239s" podCreationTimestamp="2025-12-06 04:25:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:25:48.193841427 +0000 UTC m=+1137.199546598" watchObservedRunningTime="2025-12-06 04:25:48.200047239 +0000 UTC m=+1137.205752430" Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.422192 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-vcmmj"] Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.433038 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-qkxrs"] Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.443261 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-qkxrs"] Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.453171 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-vcmmj"] Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.460165 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-c4fd8b875-2jbxs"] Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.466199 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq"] Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.466874 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" podUID="b4c57765-c010-4794-8e32-11340a76ea67" containerName="keystone-api" containerID="cri-o://bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f" gracePeriod=30 Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.472333 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone385d-account-delete-vqtlt"] Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.473195 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.493174 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone385d-account-delete-vqtlt"] Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.505331 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc9f36c3-0de5-4372-9d70-4e2897de15ec-operator-scripts\") pod \"keystone385d-account-delete-vqtlt\" (UID: \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\") " pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.505431 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vfzq\" (UniqueName: \"kubernetes.io/projected/dc9f36c3-0de5-4372-9d70-4e2897de15ec-kube-api-access-8vfzq\") pod \"keystone385d-account-delete-vqtlt\" (UID: \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\") " pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.606225 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc9f36c3-0de5-4372-9d70-4e2897de15ec-operator-scripts\") pod \"keystone385d-account-delete-vqtlt\" (UID: \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\") " pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.606312 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vfzq\" (UniqueName: \"kubernetes.io/projected/dc9f36c3-0de5-4372-9d70-4e2897de15ec-kube-api-access-8vfzq\") pod \"keystone385d-account-delete-vqtlt\" (UID: \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\") " pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.606960 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc9f36c3-0de5-4372-9d70-4e2897de15ec-operator-scripts\") pod \"keystone385d-account-delete-vqtlt\" (UID: \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\") " pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.622019 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vfzq\" (UniqueName: \"kubernetes.io/projected/dc9f36c3-0de5-4372-9d70-4e2897de15ec-kube-api-access-8vfzq\") pod \"keystone385d-account-delete-vqtlt\" (UID: \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\") " pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:48 crc kubenswrapper[4718]: I1206 04:25:48.788416 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:49 crc kubenswrapper[4718]: I1206 04:25:49.067869 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone385d-account-delete-vqtlt"] Dec 06 04:25:49 crc kubenswrapper[4718]: W1206 04:25:49.075590 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc9f36c3_0de5_4372_9d70_4e2897de15ec.slice/crio-af14c2c15cd5caa6b75d2dd037405fc4b97036ff9bdb60889fec4706bba23f72 WatchSource:0}: Error finding container af14c2c15cd5caa6b75d2dd037405fc4b97036ff9bdb60889fec4706bba23f72: Status 404 returned error can't find the container with id af14c2c15cd5caa6b75d2dd037405fc4b97036ff9bdb60889fec4706bba23f72 Dec 06 04:25:49 crc kubenswrapper[4718]: I1206 04:25:49.171178 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" event={"ID":"dc9f36c3-0de5-4372-9d70-4e2897de15ec","Type":"ContainerStarted","Data":"af14c2c15cd5caa6b75d2dd037405fc4b97036ff9bdb60889fec4706bba23f72"} Dec 06 04:25:49 crc kubenswrapper[4718]: I1206 04:25:49.171697 4718 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" secret="" err="secret \"keystone-keystone-dockercfg-86krp\" not found" Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.316405 4718 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone-scripts: secret "keystone-scripts" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.316724 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts podName:6b518e06-ee13-41e3-8d3d-80139e9722a0 nodeName:}" failed. No retries permitted until 2025-12-06 04:25:49.816705105 +0000 UTC m=+1138.822410276 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts") pod "keystone-c4fd8b875-2jbxs" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0") : secret "keystone-scripts" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.316527 4718 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone: secret "keystone" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.316624 4718 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone: secret "keystone" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.316652 4718 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone-config-data: secret "keystone-config-data" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.316831 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys podName:6b518e06-ee13-41e3-8d3d-80139e9722a0 nodeName:}" failed. No retries permitted until 2025-12-06 04:25:49.816802418 +0000 UTC m=+1138.822507619 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "credential-keys" (UniqueName: "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys") pod "keystone-c4fd8b875-2jbxs" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0") : secret "keystone" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.316947 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys podName:6b518e06-ee13-41e3-8d3d-80139e9722a0 nodeName:}" failed. No retries permitted until 2025-12-06 04:25:49.816927391 +0000 UTC m=+1138.822632552 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "fernet-keys" (UniqueName: "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys") pod "keystone-c4fd8b875-2jbxs" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0") : secret "keystone" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.316966 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data podName:6b518e06-ee13-41e3-8d3d-80139e9722a0 nodeName:}" failed. No retries permitted until 2025-12-06 04:25:49.816959812 +0000 UTC m=+1138.822664963 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data") pod "keystone-c4fd8b875-2jbxs" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0") : secret "keystone-config-data" not found Dec 06 04:25:49 crc kubenswrapper[4718]: I1206 04:25:49.348672 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12732d17-ee22-4bef-ab33-4ef7ebd1b6f9" path="/var/lib/kubelet/pods/12732d17-ee22-4bef-ab33-4ef7ebd1b6f9/volumes" Dec 06 04:25:49 crc kubenswrapper[4718]: I1206 04:25:49.349948 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db3759d7-a87d-492b-a711-72550fc419bd" path="/var/lib/kubelet/pods/db3759d7-a87d-492b-a711-72550fc419bd/volumes" Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.824084 4718 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone: secret "keystone" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.824165 4718 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone: secret "keystone" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.824173 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys podName:6b518e06-ee13-41e3-8d3d-80139e9722a0 nodeName:}" failed. No retries permitted until 2025-12-06 04:25:50.824152721 +0000 UTC m=+1139.829857892 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "credential-keys" (UniqueName: "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys") pod "keystone-c4fd8b875-2jbxs" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0") : secret "keystone" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.824276 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys podName:6b518e06-ee13-41e3-8d3d-80139e9722a0 nodeName:}" failed. No retries permitted until 2025-12-06 04:25:50.824255494 +0000 UTC m=+1139.829960715 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "fernet-keys" (UniqueName: "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys") pod "keystone-c4fd8b875-2jbxs" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0") : secret "keystone" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.824174 4718 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone-scripts: secret "keystone-scripts" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.824299 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts podName:6b518e06-ee13-41e3-8d3d-80139e9722a0 nodeName:}" failed. No retries permitted until 2025-12-06 04:25:50.824294126 +0000 UTC m=+1139.829999287 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts") pod "keystone-c4fd8b875-2jbxs" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0") : secret "keystone-scripts" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.824174 4718 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone-config-data: secret "keystone-config-data" not found Dec 06 04:25:49 crc kubenswrapper[4718]: E1206 04:25:49.824403 4718 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data podName:6b518e06-ee13-41e3-8d3d-80139e9722a0 nodeName:}" failed. No retries permitted until 2025-12-06 04:25:50.824373928 +0000 UTC m=+1139.830079129 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data") pod "keystone-c4fd8b875-2jbxs" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0") : secret "keystone-config-data" not found Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.180712 4718 generic.go:334] "Generic (PLEG): container finished" podID="dc9f36c3-0de5-4372-9d70-4e2897de15ec" containerID="397359323d68cd5bbe95c4bd8af8bf3b69594806e2ac558f21e05cfe69cb8580" exitCode=0 Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.180783 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" event={"ID":"dc9f36c3-0de5-4372-9d70-4e2897de15ec","Type":"ContainerDied","Data":"397359323d68cd5bbe95c4bd8af8bf3b69594806e2ac558f21e05cfe69cb8580"} Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.180973 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" podUID="6b518e06-ee13-41e3-8d3d-80139e9722a0" containerName="keystone-api" containerID="cri-o://439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea" gracePeriod=30 Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.665865 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.844515 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys\") pod \"6b518e06-ee13-41e3-8d3d-80139e9722a0\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.844639 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts\") pod \"6b518e06-ee13-41e3-8d3d-80139e9722a0\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.844736 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lmjn\" (UniqueName: \"kubernetes.io/projected/6b518e06-ee13-41e3-8d3d-80139e9722a0-kube-api-access-5lmjn\") pod \"6b518e06-ee13-41e3-8d3d-80139e9722a0\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.844855 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data\") pod \"6b518e06-ee13-41e3-8d3d-80139e9722a0\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.844896 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys\") pod \"6b518e06-ee13-41e3-8d3d-80139e9722a0\" (UID: \"6b518e06-ee13-41e3-8d3d-80139e9722a0\") " Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.851443 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b518e06-ee13-41e3-8d3d-80139e9722a0-kube-api-access-5lmjn" (OuterVolumeSpecName: "kube-api-access-5lmjn") pod "6b518e06-ee13-41e3-8d3d-80139e9722a0" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0"). InnerVolumeSpecName "kube-api-access-5lmjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.851794 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6b518e06-ee13-41e3-8d3d-80139e9722a0" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.852394 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts" (OuterVolumeSpecName: "scripts") pod "6b518e06-ee13-41e3-8d3d-80139e9722a0" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.852553 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6b518e06-ee13-41e3-8d3d-80139e9722a0" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.880679 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data" (OuterVolumeSpecName: "config-data") pod "6b518e06-ee13-41e3-8d3d-80139e9722a0" (UID: "6b518e06-ee13-41e3-8d3d-80139e9722a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.946618 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lmjn\" (UniqueName: \"kubernetes.io/projected/6b518e06-ee13-41e3-8d3d-80139e9722a0-kube-api-access-5lmjn\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.946710 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.946729 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.946747 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:50 crc kubenswrapper[4718]: I1206 04:25:50.946764 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b518e06-ee13-41e3-8d3d-80139e9722a0-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.195117 4718 generic.go:334] "Generic (PLEG): container finished" podID="6b518e06-ee13-41e3-8d3d-80139e9722a0" containerID="439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea" exitCode=0 Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.195191 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.195187 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" event={"ID":"6b518e06-ee13-41e3-8d3d-80139e9722a0","Type":"ContainerDied","Data":"439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea"} Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.195298 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-c4fd8b875-2jbxs" event={"ID":"6b518e06-ee13-41e3-8d3d-80139e9722a0","Type":"ContainerDied","Data":"7c27d648c856b33fd956cf4877eda8d07d1cbad4c2b5ce7c5e5de8a6914a26f6"} Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.195332 4718 scope.go:117] "RemoveContainer" containerID="439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.219126 4718 scope.go:117] "RemoveContainer" containerID="439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea" Dec 06 04:25:51 crc kubenswrapper[4718]: E1206 04:25:51.219867 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea\": container with ID starting with 439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea not found: ID does not exist" containerID="439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.219973 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea"} err="failed to get container status \"439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea\": rpc error: code = NotFound desc = could not find container \"439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea\": container with ID starting with 439618a84dc5035dadcda181c462903c2c8d40a3ea987a5a41277606c2916dea not found: ID does not exist" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.237577 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-c4fd8b875-2jbxs"] Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.247389 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-c4fd8b875-2jbxs"] Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.343795 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b518e06-ee13-41e3-8d3d-80139e9722a0" path="/var/lib/kubelet/pods/6b518e06-ee13-41e3-8d3d-80139e9722a0/volumes" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.551124 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.656218 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vfzq\" (UniqueName: \"kubernetes.io/projected/dc9f36c3-0de5-4372-9d70-4e2897de15ec-kube-api-access-8vfzq\") pod \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\" (UID: \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\") " Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.656983 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc9f36c3-0de5-4372-9d70-4e2897de15ec-operator-scripts\") pod \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\" (UID: \"dc9f36c3-0de5-4372-9d70-4e2897de15ec\") " Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.657484 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc9f36c3-0de5-4372-9d70-4e2897de15ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc9f36c3-0de5-4372-9d70-4e2897de15ec" (UID: "dc9f36c3-0de5-4372-9d70-4e2897de15ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.660137 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc9f36c3-0de5-4372-9d70-4e2897de15ec-kube-api-access-8vfzq" (OuterVolumeSpecName: "kube-api-access-8vfzq") pod "dc9f36c3-0de5-4372-9d70-4e2897de15ec" (UID: "dc9f36c3-0de5-4372-9d70-4e2897de15ec"). InnerVolumeSpecName "kube-api-access-8vfzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.759146 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc9f36c3-0de5-4372-9d70-4e2897de15ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.759215 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vfzq\" (UniqueName: \"kubernetes.io/projected/dc9f36c3-0de5-4372-9d70-4e2897de15ec-kube-api-access-8vfzq\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.957850 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.961411 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-credential-keys\") pod \"b4c57765-c010-4794-8e32-11340a76ea67\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.961456 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-config-data\") pod \"b4c57765-c010-4794-8e32-11340a76ea67\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.961477 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-fernet-keys\") pod \"b4c57765-c010-4794-8e32-11340a76ea67\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.961496 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-scripts\") pod \"b4c57765-c010-4794-8e32-11340a76ea67\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.961532 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsbch\" (UniqueName: \"kubernetes.io/projected/b4c57765-c010-4794-8e32-11340a76ea67-kube-api-access-nsbch\") pod \"b4c57765-c010-4794-8e32-11340a76ea67\" (UID: \"b4c57765-c010-4794-8e32-11340a76ea67\") " Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.964483 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b4c57765-c010-4794-8e32-11340a76ea67" (UID: "b4c57765-c010-4794-8e32-11340a76ea67"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.965219 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-scripts" (OuterVolumeSpecName: "scripts") pod "b4c57765-c010-4794-8e32-11340a76ea67" (UID: "b4c57765-c010-4794-8e32-11340a76ea67"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.965354 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b4c57765-c010-4794-8e32-11340a76ea67" (UID: "b4c57765-c010-4794-8e32-11340a76ea67"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.965585 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4c57765-c010-4794-8e32-11340a76ea67-kube-api-access-nsbch" (OuterVolumeSpecName: "kube-api-access-nsbch") pod "b4c57765-c010-4794-8e32-11340a76ea67" (UID: "b4c57765-c010-4794-8e32-11340a76ea67"). InnerVolumeSpecName "kube-api-access-nsbch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:25:51 crc kubenswrapper[4718]: I1206 04:25:51.978076 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-config-data" (OuterVolumeSpecName: "config-data") pod "b4c57765-c010-4794-8e32-11340a76ea67" (UID: "b4c57765-c010-4794-8e32-11340a76ea67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.063463 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.063498 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.063511 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.063539 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsbch\" (UniqueName: \"kubernetes.io/projected/b4c57765-c010-4794-8e32-11340a76ea67-kube-api-access-nsbch\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.063550 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b4c57765-c010-4794-8e32-11340a76ea67-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.203417 4718 generic.go:334] "Generic (PLEG): container finished" podID="b4c57765-c010-4794-8e32-11340a76ea67" containerID="bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f" exitCode=0 Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.203474 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" event={"ID":"b4c57765-c010-4794-8e32-11340a76ea67","Type":"ContainerDied","Data":"bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f"} Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.203501 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" event={"ID":"b4c57765-c010-4794-8e32-11340a76ea67","Type":"ContainerDied","Data":"c93a9328532e55a7c200ff12728b028168edebd9b1bb416d225015be79440a34"} Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.203518 4718 scope.go:117] "RemoveContainer" containerID="bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.203590 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.207179 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" event={"ID":"dc9f36c3-0de5-4372-9d70-4e2897de15ec","Type":"ContainerDied","Data":"af14c2c15cd5caa6b75d2dd037405fc4b97036ff9bdb60889fec4706bba23f72"} Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.207208 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af14c2c15cd5caa6b75d2dd037405fc4b97036ff9bdb60889fec4706bba23f72" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.207404 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone385d-account-delete-vqtlt" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.228466 4718 scope.go:117] "RemoveContainer" containerID="bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f" Dec 06 04:25:52 crc kubenswrapper[4718]: E1206 04:25:52.229853 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f\": container with ID starting with bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f not found: ID does not exist" containerID="bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.229975 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f"} err="failed to get container status \"bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f\": rpc error: code = NotFound desc = could not find container \"bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f\": container with ID starting with bcea2864e7f4c7c32c390353c9c81ff4adbcd32e92a4437ee127bd9379fc360f not found: ID does not exist" Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.233112 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq"] Dec 06 04:25:52 crc kubenswrapper[4718]: I1206 04:25:52.237522 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-79fc8bbcc8-9wwdq"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.339360 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4c57765-c010-4794-8e32-11340a76ea67" path="/var/lib/kubelet/pods/b4c57765-c010-4794-8e32-11340a76ea67/volumes" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.510786 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-5jl2m"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.526933 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-5jl2m"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.535298 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.539463 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone385d-account-delete-vqtlt"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.543227 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-385d-account-create-update-cgvr6"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.548011 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone385d-account-delete-vqtlt"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.595985 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-zdjw8"] Dec 06 04:25:53 crc kubenswrapper[4718]: E1206 04:25:53.596313 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b518e06-ee13-41e3-8d3d-80139e9722a0" containerName="keystone-api" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.596329 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b518e06-ee13-41e3-8d3d-80139e9722a0" containerName="keystone-api" Dec 06 04:25:53 crc kubenswrapper[4718]: E1206 04:25:53.596343 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc9f36c3-0de5-4372-9d70-4e2897de15ec" containerName="mariadb-account-delete" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.596351 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc9f36c3-0de5-4372-9d70-4e2897de15ec" containerName="mariadb-account-delete" Dec 06 04:25:53 crc kubenswrapper[4718]: E1206 04:25:53.596367 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4c57765-c010-4794-8e32-11340a76ea67" containerName="keystone-api" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.596375 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4c57765-c010-4794-8e32-11340a76ea67" containerName="keystone-api" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.596513 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc9f36c3-0de5-4372-9d70-4e2897de15ec" containerName="mariadb-account-delete" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.596537 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4c57765-c010-4794-8e32-11340a76ea67" containerName="keystone-api" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.596555 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b518e06-ee13-41e3-8d3d-80139e9722a0" containerName="keystone-api" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.597155 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.645932 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-zdjw8"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.697944 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb6639e4-b470-45ca-8717-3f231badd399-operator-scripts\") pod \"keystone-db-create-zdjw8\" (UID: \"fb6639e4-b470-45ca-8717-3f231badd399\") " pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.698272 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkg64\" (UniqueName: \"kubernetes.io/projected/fb6639e4-b470-45ca-8717-3f231badd399-kube-api-access-fkg64\") pod \"keystone-db-create-zdjw8\" (UID: \"fb6639e4-b470-45ca-8717-3f231badd399\") " pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.700834 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.701979 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.703409 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.706853 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2"] Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.798955 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkg64\" (UniqueName: \"kubernetes.io/projected/fb6639e4-b470-45ca-8717-3f231badd399-kube-api-access-fkg64\") pod \"keystone-db-create-zdjw8\" (UID: \"fb6639e4-b470-45ca-8717-3f231badd399\") " pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.799009 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/183c06e8-ca44-4a1e-b6c9-c7599077370a-operator-scripts\") pod \"keystone-d058-account-create-update-m8cd2\" (UID: \"183c06e8-ca44-4a1e-b6c9-c7599077370a\") " pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.799063 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb6639e4-b470-45ca-8717-3f231badd399-operator-scripts\") pod \"keystone-db-create-zdjw8\" (UID: \"fb6639e4-b470-45ca-8717-3f231badd399\") " pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.799161 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w7zk\" (UniqueName: \"kubernetes.io/projected/183c06e8-ca44-4a1e-b6c9-c7599077370a-kube-api-access-5w7zk\") pod \"keystone-d058-account-create-update-m8cd2\" (UID: \"183c06e8-ca44-4a1e-b6c9-c7599077370a\") " pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.800891 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb6639e4-b470-45ca-8717-3f231badd399-operator-scripts\") pod \"keystone-db-create-zdjw8\" (UID: \"fb6639e4-b470-45ca-8717-3f231badd399\") " pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.823945 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkg64\" (UniqueName: \"kubernetes.io/projected/fb6639e4-b470-45ca-8717-3f231badd399-kube-api-access-fkg64\") pod \"keystone-db-create-zdjw8\" (UID: \"fb6639e4-b470-45ca-8717-3f231badd399\") " pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.899948 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w7zk\" (UniqueName: \"kubernetes.io/projected/183c06e8-ca44-4a1e-b6c9-c7599077370a-kube-api-access-5w7zk\") pod \"keystone-d058-account-create-update-m8cd2\" (UID: \"183c06e8-ca44-4a1e-b6c9-c7599077370a\") " pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.900042 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/183c06e8-ca44-4a1e-b6c9-c7599077370a-operator-scripts\") pod \"keystone-d058-account-create-update-m8cd2\" (UID: \"183c06e8-ca44-4a1e-b6c9-c7599077370a\") " pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.901007 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/183c06e8-ca44-4a1e-b6c9-c7599077370a-operator-scripts\") pod \"keystone-d058-account-create-update-m8cd2\" (UID: \"183c06e8-ca44-4a1e-b6c9-c7599077370a\") " pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.930955 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w7zk\" (UniqueName: \"kubernetes.io/projected/183c06e8-ca44-4a1e-b6c9-c7599077370a-kube-api-access-5w7zk\") pod \"keystone-d058-account-create-update-m8cd2\" (UID: \"183c06e8-ca44-4a1e-b6c9-c7599077370a\") " pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:53 crc kubenswrapper[4718]: I1206 04:25:53.958569 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:54 crc kubenswrapper[4718]: I1206 04:25:54.021012 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:54 crc kubenswrapper[4718]: I1206 04:25:54.418721 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-zdjw8"] Dec 06 04:25:54 crc kubenswrapper[4718]: W1206 04:25:54.422101 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb6639e4_b470_45ca_8717_3f231badd399.slice/crio-aee377547c4fb89604c75463f82318e9f87d5bc4143db016cb888de1a08767c3 WatchSource:0}: Error finding container aee377547c4fb89604c75463f82318e9f87d5bc4143db016cb888de1a08767c3: Status 404 returned error can't find the container with id aee377547c4fb89604c75463f82318e9f87d5bc4143db016cb888de1a08767c3 Dec 06 04:25:54 crc kubenswrapper[4718]: I1206 04:25:54.494816 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2"] Dec 06 04:25:54 crc kubenswrapper[4718]: W1206 04:25:54.501484 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod183c06e8_ca44_4a1e_b6c9_c7599077370a.slice/crio-9f41427c9e4e8e5d413901f1bfe218beb7582d7b1d6b2be9aaf16f81db21dfcf WatchSource:0}: Error finding container 9f41427c9e4e8e5d413901f1bfe218beb7582d7b1d6b2be9aaf16f81db21dfcf: Status 404 returned error can't find the container with id 9f41427c9e4e8e5d413901f1bfe218beb7582d7b1d6b2be9aaf16f81db21dfcf Dec 06 04:25:55 crc kubenswrapper[4718]: I1206 04:25:55.233090 4718 generic.go:334] "Generic (PLEG): container finished" podID="fb6639e4-b470-45ca-8717-3f231badd399" containerID="210b717444b8d92e2296cad127fce92445392521e154df5f6ef865bc21984a4f" exitCode=0 Dec 06 04:25:55 crc kubenswrapper[4718]: I1206 04:25:55.233357 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-zdjw8" event={"ID":"fb6639e4-b470-45ca-8717-3f231badd399","Type":"ContainerDied","Data":"210b717444b8d92e2296cad127fce92445392521e154df5f6ef865bc21984a4f"} Dec 06 04:25:55 crc kubenswrapper[4718]: I1206 04:25:55.233461 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-zdjw8" event={"ID":"fb6639e4-b470-45ca-8717-3f231badd399","Type":"ContainerStarted","Data":"aee377547c4fb89604c75463f82318e9f87d5bc4143db016cb888de1a08767c3"} Dec 06 04:25:55 crc kubenswrapper[4718]: I1206 04:25:55.239630 4718 generic.go:334] "Generic (PLEG): container finished" podID="183c06e8-ca44-4a1e-b6c9-c7599077370a" containerID="93b6d25531f6f3944886d0e3bc606d3691ad13d1f0d63b31160049fd25dc9470" exitCode=0 Dec 06 04:25:55 crc kubenswrapper[4718]: I1206 04:25:55.239682 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" event={"ID":"183c06e8-ca44-4a1e-b6c9-c7599077370a","Type":"ContainerDied","Data":"93b6d25531f6f3944886d0e3bc606d3691ad13d1f0d63b31160049fd25dc9470"} Dec 06 04:25:55 crc kubenswrapper[4718]: I1206 04:25:55.239720 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" event={"ID":"183c06e8-ca44-4a1e-b6c9-c7599077370a","Type":"ContainerStarted","Data":"9f41427c9e4e8e5d413901f1bfe218beb7582d7b1d6b2be9aaf16f81db21dfcf"} Dec 06 04:25:55 crc kubenswrapper[4718]: I1206 04:25:55.344099 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65d645a2-187b-4fb8-af34-5d1bad8d580b" path="/var/lib/kubelet/pods/65d645a2-187b-4fb8-af34-5d1bad8d580b/volumes" Dec 06 04:25:55 crc kubenswrapper[4718]: I1206 04:25:55.345199 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc9f36c3-0de5-4372-9d70-4e2897de15ec" path="/var/lib/kubelet/pods/dc9f36c3-0de5-4372-9d70-4e2897de15ec/volumes" Dec 06 04:25:55 crc kubenswrapper[4718]: I1206 04:25:55.346144 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4430b20-c353-47e4-9fb3-ea1aaeb35a16" path="/var/lib/kubelet/pods/e4430b20-c353-47e4-9fb3-ea1aaeb35a16/volumes" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.618021 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.623178 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.654086 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb6639e4-b470-45ca-8717-3f231badd399-operator-scripts\") pod \"fb6639e4-b470-45ca-8717-3f231badd399\" (UID: \"fb6639e4-b470-45ca-8717-3f231badd399\") " Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.654482 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/183c06e8-ca44-4a1e-b6c9-c7599077370a-operator-scripts\") pod \"183c06e8-ca44-4a1e-b6c9-c7599077370a\" (UID: \"183c06e8-ca44-4a1e-b6c9-c7599077370a\") " Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.654672 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkg64\" (UniqueName: \"kubernetes.io/projected/fb6639e4-b470-45ca-8717-3f231badd399-kube-api-access-fkg64\") pod \"fb6639e4-b470-45ca-8717-3f231badd399\" (UID: \"fb6639e4-b470-45ca-8717-3f231badd399\") " Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.654912 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w7zk\" (UniqueName: \"kubernetes.io/projected/183c06e8-ca44-4a1e-b6c9-c7599077370a-kube-api-access-5w7zk\") pod \"183c06e8-ca44-4a1e-b6c9-c7599077370a\" (UID: \"183c06e8-ca44-4a1e-b6c9-c7599077370a\") " Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.655057 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb6639e4-b470-45ca-8717-3f231badd399-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fb6639e4-b470-45ca-8717-3f231badd399" (UID: "fb6639e4-b470-45ca-8717-3f231badd399"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.655127 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/183c06e8-ca44-4a1e-b6c9-c7599077370a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "183c06e8-ca44-4a1e-b6c9-c7599077370a" (UID: "183c06e8-ca44-4a1e-b6c9-c7599077370a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.655480 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb6639e4-b470-45ca-8717-3f231badd399-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.655570 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/183c06e8-ca44-4a1e-b6c9-c7599077370a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.661005 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb6639e4-b470-45ca-8717-3f231badd399-kube-api-access-fkg64" (OuterVolumeSpecName: "kube-api-access-fkg64") pod "fb6639e4-b470-45ca-8717-3f231badd399" (UID: "fb6639e4-b470-45ca-8717-3f231badd399"). InnerVolumeSpecName "kube-api-access-fkg64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.661102 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/183c06e8-ca44-4a1e-b6c9-c7599077370a-kube-api-access-5w7zk" (OuterVolumeSpecName: "kube-api-access-5w7zk") pod "183c06e8-ca44-4a1e-b6c9-c7599077370a" (UID: "183c06e8-ca44-4a1e-b6c9-c7599077370a"). InnerVolumeSpecName "kube-api-access-5w7zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.756792 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkg64\" (UniqueName: \"kubernetes.io/projected/fb6639e4-b470-45ca-8717-3f231badd399-kube-api-access-fkg64\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:56 crc kubenswrapper[4718]: I1206 04:25:56.756836 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w7zk\" (UniqueName: \"kubernetes.io/projected/183c06e8-ca44-4a1e-b6c9-c7599077370a-kube-api-access-5w7zk\") on node \"crc\" DevicePath \"\"" Dec 06 04:25:57 crc kubenswrapper[4718]: I1206 04:25:57.262853 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-zdjw8" event={"ID":"fb6639e4-b470-45ca-8717-3f231badd399","Type":"ContainerDied","Data":"aee377547c4fb89604c75463f82318e9f87d5bc4143db016cb888de1a08767c3"} Dec 06 04:25:57 crc kubenswrapper[4718]: I1206 04:25:57.262895 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aee377547c4fb89604c75463f82318e9f87d5bc4143db016cb888de1a08767c3" Dec 06 04:25:57 crc kubenswrapper[4718]: I1206 04:25:57.262868 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-zdjw8" Dec 06 04:25:57 crc kubenswrapper[4718]: I1206 04:25:57.264403 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" event={"ID":"183c06e8-ca44-4a1e-b6c9-c7599077370a","Type":"ContainerDied","Data":"9f41427c9e4e8e5d413901f1bfe218beb7582d7b1d6b2be9aaf16f81db21dfcf"} Dec 06 04:25:57 crc kubenswrapper[4718]: I1206 04:25:57.264426 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f41427c9e4e8e5d413901f1bfe218beb7582d7b1d6b2be9aaf16f81db21dfcf" Dec 06 04:25:57 crc kubenswrapper[4718]: I1206 04:25:57.264472 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.152495 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-v5nl4"] Dec 06 04:25:59 crc kubenswrapper[4718]: E1206 04:25:59.153664 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb6639e4-b470-45ca-8717-3f231badd399" containerName="mariadb-database-create" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.153763 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb6639e4-b470-45ca-8717-3f231badd399" containerName="mariadb-database-create" Dec 06 04:25:59 crc kubenswrapper[4718]: E1206 04:25:59.153874 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="183c06e8-ca44-4a1e-b6c9-c7599077370a" containerName="mariadb-account-create-update" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.153947 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="183c06e8-ca44-4a1e-b6c9-c7599077370a" containerName="mariadb-account-create-update" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.154171 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="183c06e8-ca44-4a1e-b6c9-c7599077370a" containerName="mariadb-account-create-update" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.154292 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb6639e4-b470-45ca-8717-3f231badd399" containerName="mariadb-database-create" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.154872 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.157178 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-vpcht" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.157520 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.157693 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.158076 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.163666 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-v5nl4"] Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.190436 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-config-data\") pod \"keystone-db-sync-v5nl4\" (UID: \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\") " pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.190611 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kg4f\" (UniqueName: \"kubernetes.io/projected/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-kube-api-access-6kg4f\") pod \"keystone-db-sync-v5nl4\" (UID: \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\") " pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.292177 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-config-data\") pod \"keystone-db-sync-v5nl4\" (UID: \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\") " pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.292332 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kg4f\" (UniqueName: \"kubernetes.io/projected/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-kube-api-access-6kg4f\") pod \"keystone-db-sync-v5nl4\" (UID: \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\") " pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.298482 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-config-data\") pod \"keystone-db-sync-v5nl4\" (UID: \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\") " pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.328078 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kg4f\" (UniqueName: \"kubernetes.io/projected/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-kube-api-access-6kg4f\") pod \"keystone-db-sync-v5nl4\" (UID: \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\") " pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.484638 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:25:59 crc kubenswrapper[4718]: I1206 04:25:59.757382 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-v5nl4"] Dec 06 04:26:00 crc kubenswrapper[4718]: I1206 04:26:00.288588 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" event={"ID":"2f0aadfe-a9c3-4c89-be50-e91df227c3a2","Type":"ContainerStarted","Data":"154af6f90e3efe9d74b39726a5b859e7935e42a7499c10b27f7046c0e470d9f0"} Dec 06 04:26:00 crc kubenswrapper[4718]: I1206 04:26:00.288922 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" event={"ID":"2f0aadfe-a9c3-4c89-be50-e91df227c3a2","Type":"ContainerStarted","Data":"ee635ae304b859646742825321528c44f9e0d4dcee08766bb8b6b8903bac692f"} Dec 06 04:26:02 crc kubenswrapper[4718]: I1206 04:26:02.330517 4718 generic.go:334] "Generic (PLEG): container finished" podID="2f0aadfe-a9c3-4c89-be50-e91df227c3a2" containerID="154af6f90e3efe9d74b39726a5b859e7935e42a7499c10b27f7046c0e470d9f0" exitCode=0 Dec 06 04:26:02 crc kubenswrapper[4718]: I1206 04:26:02.330609 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" event={"ID":"2f0aadfe-a9c3-4c89-be50-e91df227c3a2","Type":"ContainerDied","Data":"154af6f90e3efe9d74b39726a5b859e7935e42a7499c10b27f7046c0e470d9f0"} Dec 06 04:26:03 crc kubenswrapper[4718]: I1206 04:26:03.663785 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:26:03 crc kubenswrapper[4718]: I1206 04:26:03.756834 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kg4f\" (UniqueName: \"kubernetes.io/projected/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-kube-api-access-6kg4f\") pod \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\" (UID: \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\") " Dec 06 04:26:03 crc kubenswrapper[4718]: I1206 04:26:03.756942 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-config-data\") pod \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\" (UID: \"2f0aadfe-a9c3-4c89-be50-e91df227c3a2\") " Dec 06 04:26:03 crc kubenswrapper[4718]: I1206 04:26:03.762920 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-kube-api-access-6kg4f" (OuterVolumeSpecName: "kube-api-access-6kg4f") pod "2f0aadfe-a9c3-4c89-be50-e91df227c3a2" (UID: "2f0aadfe-a9c3-4c89-be50-e91df227c3a2"). InnerVolumeSpecName "kube-api-access-6kg4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:26:03 crc kubenswrapper[4718]: I1206 04:26:03.821072 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-config-data" (OuterVolumeSpecName: "config-data") pod "2f0aadfe-a9c3-4c89-be50-e91df227c3a2" (UID: "2f0aadfe-a9c3-4c89-be50-e91df227c3a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:26:03 crc kubenswrapper[4718]: I1206 04:26:03.866563 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:26:03 crc kubenswrapper[4718]: I1206 04:26:03.866609 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kg4f\" (UniqueName: \"kubernetes.io/projected/2f0aadfe-a9c3-4c89-be50-e91df227c3a2-kube-api-access-6kg4f\") on node \"crc\" DevicePath \"\"" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.349678 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" event={"ID":"2f0aadfe-a9c3-4c89-be50-e91df227c3a2","Type":"ContainerDied","Data":"ee635ae304b859646742825321528c44f9e0d4dcee08766bb8b6b8903bac692f"} Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.349746 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee635ae304b859646742825321528c44f9e0d4dcee08766bb8b6b8903bac692f" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.349763 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-v5nl4" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.553644 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-s2j8g"] Dec 06 04:26:04 crc kubenswrapper[4718]: E1206 04:26:04.554029 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0aadfe-a9c3-4c89-be50-e91df227c3a2" containerName="keystone-db-sync" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.554054 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0aadfe-a9c3-4c89-be50-e91df227c3a2" containerName="keystone-db-sync" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.554261 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0aadfe-a9c3-4c89-be50-e91df227c3a2" containerName="keystone-db-sync" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.554949 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.559595 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-vpcht" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.560285 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.560337 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.560344 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.563464 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.574023 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-s2j8g"] Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.679737 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-scripts\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.679792 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-config-data\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.679831 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xqr5\" (UniqueName: \"kubernetes.io/projected/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-kube-api-access-7xqr5\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.680003 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-fernet-keys\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.680047 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-credential-keys\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.781466 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-scripts\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.781537 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-config-data\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.781575 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xqr5\" (UniqueName: \"kubernetes.io/projected/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-kube-api-access-7xqr5\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.781607 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-fernet-keys\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.781627 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-credential-keys\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.788923 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-credential-keys\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.789129 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-scripts\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.790417 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-config-data\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.790487 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-fernet-keys\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.815284 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xqr5\" (UniqueName: \"kubernetes.io/projected/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-kube-api-access-7xqr5\") pod \"keystone-bootstrap-s2j8g\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:04 crc kubenswrapper[4718]: I1206 04:26:04.878077 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:05 crc kubenswrapper[4718]: I1206 04:26:05.123536 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-s2j8g"] Dec 06 04:26:05 crc kubenswrapper[4718]: I1206 04:26:05.359626 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" event={"ID":"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9","Type":"ContainerStarted","Data":"eadff7670ab04352e8375e280cc4b03c9fa9ce1d28f8dedd5455a91a29f98216"} Dec 06 04:26:06 crc kubenswrapper[4718]: I1206 04:26:06.367774 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" event={"ID":"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9","Type":"ContainerStarted","Data":"d9b39207b2f4dcb5f75c163d4bb0aca0dec7dcb0fee345cd7d2d1c7cf677674a"} Dec 06 04:26:06 crc kubenswrapper[4718]: I1206 04:26:06.389924 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" podStartSLOduration=2.389896272 podStartE2EDuration="2.389896272s" podCreationTimestamp="2025-12-06 04:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:26:06.386190429 +0000 UTC m=+1155.391895650" watchObservedRunningTime="2025-12-06 04:26:06.389896272 +0000 UTC m=+1155.395601463" Dec 06 04:26:08 crc kubenswrapper[4718]: I1206 04:26:08.379667 4718 generic.go:334] "Generic (PLEG): container finished" podID="b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" containerID="d9b39207b2f4dcb5f75c163d4bb0aca0dec7dcb0fee345cd7d2d1c7cf677674a" exitCode=0 Dec 06 04:26:08 crc kubenswrapper[4718]: I1206 04:26:08.379803 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" event={"ID":"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9","Type":"ContainerDied","Data":"d9b39207b2f4dcb5f75c163d4bb0aca0dec7dcb0fee345cd7d2d1c7cf677674a"} Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.736363 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.866498 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-config-data\") pod \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.866556 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-credential-keys\") pod \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.866606 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xqr5\" (UniqueName: \"kubernetes.io/projected/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-kube-api-access-7xqr5\") pod \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.866639 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-fernet-keys\") pod \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.866704 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-scripts\") pod \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\" (UID: \"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9\") " Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.872910 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-scripts" (OuterVolumeSpecName: "scripts") pod "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" (UID: "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.873935 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" (UID: "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.874083 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-kube-api-access-7xqr5" (OuterVolumeSpecName: "kube-api-access-7xqr5") pod "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" (UID: "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9"). InnerVolumeSpecName "kube-api-access-7xqr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.874832 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" (UID: "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.901907 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-config-data" (OuterVolumeSpecName: "config-data") pod "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" (UID: "b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.968978 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.969036 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.969062 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xqr5\" (UniqueName: \"kubernetes.io/projected/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-kube-api-access-7xqr5\") on node \"crc\" DevicePath \"\"" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.969082 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:26:09 crc kubenswrapper[4718]: I1206 04:26:09.969101 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.400656 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" event={"ID":"b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9","Type":"ContainerDied","Data":"eadff7670ab04352e8375e280cc4b03c9fa9ce1d28f8dedd5455a91a29f98216"} Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.400982 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eadff7670ab04352e8375e280cc4b03c9fa9ce1d28f8dedd5455a91a29f98216" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.400775 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-s2j8g" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.583035 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-p6686"] Dec 06 04:26:10 crc kubenswrapper[4718]: E1206 04:26:10.583372 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" containerName="keystone-bootstrap" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.583422 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" containerName="keystone-bootstrap" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.583571 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" containerName="keystone-bootstrap" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.584062 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.587739 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.587739 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-vpcht" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.588059 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.598644 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.607969 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-p6686"] Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.679089 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-fernet-keys\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.679209 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kmwr\" (UniqueName: \"kubernetes.io/projected/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-kube-api-access-4kmwr\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.679262 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-credential-keys\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.679282 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-config-data\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.679325 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-scripts\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.780409 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kmwr\" (UniqueName: \"kubernetes.io/projected/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-kube-api-access-4kmwr\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.780503 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-credential-keys\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.780534 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-config-data\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.780555 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-scripts\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.780612 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-fernet-keys\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.787054 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-credential-keys\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.787471 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-config-data\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.790582 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-fernet-keys\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.793417 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-scripts\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.814427 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kmwr\" (UniqueName: \"kubernetes.io/projected/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-kube-api-access-4kmwr\") pod \"keystone-7db797d66-p6686\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:10 crc kubenswrapper[4718]: I1206 04:26:10.922609 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:11 crc kubenswrapper[4718]: I1206 04:26:11.385006 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-p6686"] Dec 06 04:26:11 crc kubenswrapper[4718]: W1206 04:26:11.398673 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4df8fbb0_37c3_4acb_99fe_1d07364af0b9.slice/crio-6c7ecce0be6e57e5864357b23985048174eb51485e690fc3946b63fbccfa11e3 WatchSource:0}: Error finding container 6c7ecce0be6e57e5864357b23985048174eb51485e690fc3946b63fbccfa11e3: Status 404 returned error can't find the container with id 6c7ecce0be6e57e5864357b23985048174eb51485e690fc3946b63fbccfa11e3 Dec 06 04:26:11 crc kubenswrapper[4718]: I1206 04:26:11.418388 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" event={"ID":"4df8fbb0-37c3-4acb-99fe-1d07364af0b9","Type":"ContainerStarted","Data":"6c7ecce0be6e57e5864357b23985048174eb51485e690fc3946b63fbccfa11e3"} Dec 06 04:26:13 crc kubenswrapper[4718]: I1206 04:26:13.444858 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" event={"ID":"4df8fbb0-37c3-4acb-99fe-1d07364af0b9","Type":"ContainerStarted","Data":"59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243"} Dec 06 04:26:13 crc kubenswrapper[4718]: I1206 04:26:13.445424 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:13 crc kubenswrapper[4718]: I1206 04:26:13.470792 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" podStartSLOduration=3.470764334 podStartE2EDuration="3.470764334s" podCreationTimestamp="2025-12-06 04:26:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:26:13.466988959 +0000 UTC m=+1162.472694150" watchObservedRunningTime="2025-12-06 04:26:13.470764334 +0000 UTC m=+1162.476469535" Dec 06 04:26:42 crc kubenswrapper[4718]: I1206 04:26:42.310190 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.396911 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-fcjtq"] Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.397965 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.411654 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-f2sv5"] Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.412662 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.418068 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-fcjtq"] Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.426614 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-f2sv5"] Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512101 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-credential-keys\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512138 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-scripts\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512252 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzpmf\" (UniqueName: \"kubernetes.io/projected/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-kube-api-access-mzpmf\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512312 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-fernet-keys\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512362 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-credential-keys\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512543 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-scripts\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512574 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-config-data\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512599 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-config-data\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512646 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-fernet-keys\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.512678 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zcwk\" (UniqueName: \"kubernetes.io/projected/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-kube-api-access-9zcwk\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614284 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-fernet-keys\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614347 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-credential-keys\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614435 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-scripts\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614497 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-config-data\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614524 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-config-data\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614554 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-fernet-keys\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614582 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zcwk\" (UniqueName: \"kubernetes.io/projected/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-kube-api-access-9zcwk\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614619 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-credential-keys\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614642 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-scripts\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.614664 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzpmf\" (UniqueName: \"kubernetes.io/projected/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-kube-api-access-mzpmf\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.621726 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-credential-keys\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.623004 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-credential-keys\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.623106 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-config-data\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.623099 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-scripts\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.623314 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-fernet-keys\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.630817 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-fernet-keys\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.632196 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-scripts\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.632560 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-config-data\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.635289 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zcwk\" (UniqueName: \"kubernetes.io/projected/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-kube-api-access-9zcwk\") pod \"keystone-7db797d66-fcjtq\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.645283 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzpmf\" (UniqueName: \"kubernetes.io/projected/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-kube-api-access-mzpmf\") pod \"keystone-7db797d66-f2sv5\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.720376 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:43 crc kubenswrapper[4718]: I1206 04:26:43.733681 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.148347 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-fcjtq"] Dec 06 04:26:44 crc kubenswrapper[4718]: W1206 04:26:44.157951 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44f9c92a_dd8e_43e6_a9af_9235ff6e210f.slice/crio-3c4a5f4de7ced566193c5ba9d7c66c999cef93abdf35c25a8b3c982a61f5e275 WatchSource:0}: Error finding container 3c4a5f4de7ced566193c5ba9d7c66c999cef93abdf35c25a8b3c982a61f5e275: Status 404 returned error can't find the container with id 3c4a5f4de7ced566193c5ba9d7c66c999cef93abdf35c25a8b3c982a61f5e275 Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.203983 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-f2sv5"] Dec 06 04:26:44 crc kubenswrapper[4718]: W1206 04:26:44.215990 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bd58f4f_aa4d_47e0_a2a8_cabce0e09569.slice/crio-aa8563c977d0efde837c8da8d03a8613c360165cbcd32641eca534a45ad2a141 WatchSource:0}: Error finding container aa8563c977d0efde837c8da8d03a8613c360165cbcd32641eca534a45ad2a141: Status 404 returned error can't find the container with id aa8563c977d0efde837c8da8d03a8613c360165cbcd32641eca534a45ad2a141 Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.687665 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" event={"ID":"44f9c92a-dd8e-43e6-a9af-9235ff6e210f","Type":"ContainerStarted","Data":"8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82"} Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.687923 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" event={"ID":"44f9c92a-dd8e-43e6-a9af-9235ff6e210f","Type":"ContainerStarted","Data":"3c4a5f4de7ced566193c5ba9d7c66c999cef93abdf35c25a8b3c982a61f5e275"} Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.687940 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.690391 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" event={"ID":"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569","Type":"ContainerStarted","Data":"df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695"} Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.690432 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" event={"ID":"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569","Type":"ContainerStarted","Data":"aa8563c977d0efde837c8da8d03a8613c360165cbcd32641eca534a45ad2a141"} Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.690560 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.708537 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" podStartSLOduration=1.708521863 podStartE2EDuration="1.708521863s" podCreationTimestamp="2025-12-06 04:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:26:44.703893935 +0000 UTC m=+1193.709599096" watchObservedRunningTime="2025-12-06 04:26:44.708521863 +0000 UTC m=+1193.714227024" Dec 06 04:26:44 crc kubenswrapper[4718]: I1206 04:26:44.720650 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" podStartSLOduration=1.720635449 podStartE2EDuration="1.720635449s" podCreationTimestamp="2025-12-06 04:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:26:44.71885459 +0000 UTC m=+1193.724559741" watchObservedRunningTime="2025-12-06 04:26:44.720635449 +0000 UTC m=+1193.726340610" Dec 06 04:26:57 crc kubenswrapper[4718]: I1206 04:26:57.876907 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:26:57 crc kubenswrapper[4718]: I1206 04:26:57.877584 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:27:15 crc kubenswrapper[4718]: I1206 04:27:15.166434 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:27:15 crc kubenswrapper[4718]: I1206 04:27:15.289224 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:27:16 crc kubenswrapper[4718]: I1206 04:27:16.271622 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-f2sv5"] Dec 06 04:27:16 crc kubenswrapper[4718]: I1206 04:27:16.271972 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" podUID="2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" containerName="keystone-api" containerID="cri-o://df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695" gracePeriod=30 Dec 06 04:27:16 crc kubenswrapper[4718]: I1206 04:27:16.279723 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-fcjtq"] Dec 06 04:27:16 crc kubenswrapper[4718]: I1206 04:27:16.280272 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" podUID="44f9c92a-dd8e-43e6-a9af-9235ff6e210f" containerName="keystone-api" containerID="cri-o://8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82" gracePeriod=30 Dec 06 04:27:17 crc kubenswrapper[4718]: I1206 04:27:17.477084 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-p6686"] Dec 06 04:27:17 crc kubenswrapper[4718]: I1206 04:27:17.477371 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" podUID="4df8fbb0-37c3-4acb-99fe-1d07364af0b9" containerName="keystone-api" containerID="cri-o://59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243" gracePeriod=30 Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.797594 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.803296 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.929652 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zcwk\" (UniqueName: \"kubernetes.io/projected/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-kube-api-access-9zcwk\") pod \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.929709 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzpmf\" (UniqueName: \"kubernetes.io/projected/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-kube-api-access-mzpmf\") pod \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.929773 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-fernet-keys\") pod \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.929808 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-scripts\") pod \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.931028 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-credential-keys\") pod \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.931055 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-credential-keys\") pod \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.931116 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-scripts\") pod \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.931142 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-config-data\") pod \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\" (UID: \"44f9c92a-dd8e-43e6-a9af-9235ff6e210f\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.931165 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-fernet-keys\") pod \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.931194 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-config-data\") pod \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\" (UID: \"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569\") " Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.936324 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "44f9c92a-dd8e-43e6-a9af-9235ff6e210f" (UID: "44f9c92a-dd8e-43e6-a9af-9235ff6e210f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.936555 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-scripts" (OuterVolumeSpecName: "scripts") pod "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" (UID: "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.936727 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" (UID: "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.937028 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-kube-api-access-9zcwk" (OuterVolumeSpecName: "kube-api-access-9zcwk") pod "44f9c92a-dd8e-43e6-a9af-9235ff6e210f" (UID: "44f9c92a-dd8e-43e6-a9af-9235ff6e210f"). InnerVolumeSpecName "kube-api-access-9zcwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.937149 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "44f9c92a-dd8e-43e6-a9af-9235ff6e210f" (UID: "44f9c92a-dd8e-43e6-a9af-9235ff6e210f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.937349 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-scripts" (OuterVolumeSpecName: "scripts") pod "44f9c92a-dd8e-43e6-a9af-9235ff6e210f" (UID: "44f9c92a-dd8e-43e6-a9af-9235ff6e210f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.937663 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" (UID: "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.947518 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-kube-api-access-mzpmf" (OuterVolumeSpecName: "kube-api-access-mzpmf") pod "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" (UID: "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569"). InnerVolumeSpecName "kube-api-access-mzpmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.952764 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-config-data" (OuterVolumeSpecName: "config-data") pod "44f9c92a-dd8e-43e6-a9af-9235ff6e210f" (UID: "44f9c92a-dd8e-43e6-a9af-9235ff6e210f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:19 crc kubenswrapper[4718]: I1206 04:27:19.963227 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-config-data" (OuterVolumeSpecName: "config-data") pod "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" (UID: "2bd58f4f-aa4d-47e0-a2a8-cabce0e09569"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.016828 4718 generic.go:334] "Generic (PLEG): container finished" podID="44f9c92a-dd8e-43e6-a9af-9235ff6e210f" containerID="8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82" exitCode=0 Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.016913 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.016924 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" event={"ID":"44f9c92a-dd8e-43e6-a9af-9235ff6e210f","Type":"ContainerDied","Data":"8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82"} Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.016968 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-fcjtq" event={"ID":"44f9c92a-dd8e-43e6-a9af-9235ff6e210f","Type":"ContainerDied","Data":"3c4a5f4de7ced566193c5ba9d7c66c999cef93abdf35c25a8b3c982a61f5e275"} Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.016995 4718 scope.go:117] "RemoveContainer" containerID="8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.020803 4718 generic.go:334] "Generic (PLEG): container finished" podID="2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" containerID="df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695" exitCode=0 Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.020834 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" event={"ID":"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569","Type":"ContainerDied","Data":"df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695"} Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.020855 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" event={"ID":"2bd58f4f-aa4d-47e0-a2a8-cabce0e09569","Type":"ContainerDied","Data":"aa8563c977d0efde837c8da8d03a8613c360165cbcd32641eca534a45ad2a141"} Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.020854 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-f2sv5" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033131 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zcwk\" (UniqueName: \"kubernetes.io/projected/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-kube-api-access-9zcwk\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033166 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzpmf\" (UniqueName: \"kubernetes.io/projected/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-kube-api-access-mzpmf\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033181 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033196 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033209 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033221 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033257 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033272 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44f9c92a-dd8e-43e6-a9af-9235ff6e210f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033314 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.033325 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.055799 4718 scope.go:117] "RemoveContainer" containerID="8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.061362 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-f2sv5"] Dec 06 04:27:20 crc kubenswrapper[4718]: E1206 04:27:20.062116 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82\": container with ID starting with 8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82 not found: ID does not exist" containerID="8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.062164 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82"} err="failed to get container status \"8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82\": rpc error: code = NotFound desc = could not find container \"8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82\": container with ID starting with 8cd3eb6d1f7a4138802c250c1ce3808dfd8694b15e3deddca5732885e63b9e82 not found: ID does not exist" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.062195 4718 scope.go:117] "RemoveContainer" containerID="df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.071266 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-f2sv5"] Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.078532 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-fcjtq"] Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.089328 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-fcjtq"] Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.091747 4718 scope.go:117] "RemoveContainer" containerID="df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695" Dec 06 04:27:20 crc kubenswrapper[4718]: E1206 04:27:20.092171 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695\": container with ID starting with df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695 not found: ID does not exist" containerID="df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.092217 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695"} err="failed to get container status \"df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695\": rpc error: code = NotFound desc = could not find container \"df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695\": container with ID starting with df2d95bcf44f7c73d81613479db7142092ae0b5cf7f8e08403a582a0d6650695 not found: ID does not exist" Dec 06 04:27:20 crc kubenswrapper[4718]: I1206 04:27:20.991884 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.029981 4718 generic.go:334] "Generic (PLEG): container finished" podID="4df8fbb0-37c3-4acb-99fe-1d07364af0b9" containerID="59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243" exitCode=0 Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.030066 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" event={"ID":"4df8fbb0-37c3-4acb-99fe-1d07364af0b9","Type":"ContainerDied","Data":"59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243"} Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.030078 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.030092 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7db797d66-p6686" event={"ID":"4df8fbb0-37c3-4acb-99fe-1d07364af0b9","Type":"ContainerDied","Data":"6c7ecce0be6e57e5864357b23985048174eb51485e690fc3946b63fbccfa11e3"} Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.030109 4718 scope.go:117] "RemoveContainer" containerID="59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.049089 4718 scope.go:117] "RemoveContainer" containerID="59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243" Dec 06 04:27:21 crc kubenswrapper[4718]: E1206 04:27:21.051410 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243\": container with ID starting with 59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243 not found: ID does not exist" containerID="59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.051454 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243"} err="failed to get container status \"59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243\": rpc error: code = NotFound desc = could not find container \"59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243\": container with ID starting with 59b973e270be107f527c5d7b1127fa855c2ac105edaee0530bc4a8fc6c9e1243 not found: ID does not exist" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.152652 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-credential-keys\") pod \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.152726 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-fernet-keys\") pod \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.152767 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kmwr\" (UniqueName: \"kubernetes.io/projected/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-kube-api-access-4kmwr\") pod \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.152797 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-scripts\") pod \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.152814 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-config-data\") pod \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\" (UID: \"4df8fbb0-37c3-4acb-99fe-1d07364af0b9\") " Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.157527 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4df8fbb0-37c3-4acb-99fe-1d07364af0b9" (UID: "4df8fbb0-37c3-4acb-99fe-1d07364af0b9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.157856 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4df8fbb0-37c3-4acb-99fe-1d07364af0b9" (UID: "4df8fbb0-37c3-4acb-99fe-1d07364af0b9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.158391 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-kube-api-access-4kmwr" (OuterVolumeSpecName: "kube-api-access-4kmwr") pod "4df8fbb0-37c3-4acb-99fe-1d07364af0b9" (UID: "4df8fbb0-37c3-4acb-99fe-1d07364af0b9"). InnerVolumeSpecName "kube-api-access-4kmwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.158789 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-scripts" (OuterVolumeSpecName: "scripts") pod "4df8fbb0-37c3-4acb-99fe-1d07364af0b9" (UID: "4df8fbb0-37c3-4acb-99fe-1d07364af0b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.174445 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-config-data" (OuterVolumeSpecName: "config-data") pod "4df8fbb0-37c3-4acb-99fe-1d07364af0b9" (UID: "4df8fbb0-37c3-4acb-99fe-1d07364af0b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.253998 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.254027 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.254037 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kmwr\" (UniqueName: \"kubernetes.io/projected/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-kube-api-access-4kmwr\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.254047 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.254055 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df8fbb0-37c3-4acb-99fe-1d07364af0b9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.342797 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" path="/var/lib/kubelet/pods/2bd58f4f-aa4d-47e0-a2a8-cabce0e09569/volumes" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.343383 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44f9c92a-dd8e-43e6-a9af-9235ff6e210f" path="/var/lib/kubelet/pods/44f9c92a-dd8e-43e6-a9af-9235ff6e210f/volumes" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.389131 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-p6686"] Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.395559 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-7db797d66-p6686"] Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.610731 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-s2j8g"] Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.611916 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-s2j8g"] Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.617648 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-v5nl4"] Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.625330 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-v5nl4"] Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.671703 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystoned058-account-delete-kl8ff"] Dec 06 04:27:21 crc kubenswrapper[4718]: E1206 04:27:21.671930 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" containerName="keystone-api" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.671972 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" containerName="keystone-api" Dec 06 04:27:21 crc kubenswrapper[4718]: E1206 04:27:21.671984 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44f9c92a-dd8e-43e6-a9af-9235ff6e210f" containerName="keystone-api" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.671990 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="44f9c92a-dd8e-43e6-a9af-9235ff6e210f" containerName="keystone-api" Dec 06 04:27:21 crc kubenswrapper[4718]: E1206 04:27:21.672002 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4df8fbb0-37c3-4acb-99fe-1d07364af0b9" containerName="keystone-api" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.672007 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="4df8fbb0-37c3-4acb-99fe-1d07364af0b9" containerName="keystone-api" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.672107 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="4df8fbb0-37c3-4acb-99fe-1d07364af0b9" containerName="keystone-api" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.672119 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd58f4f-aa4d-47e0-a2a8-cabce0e09569" containerName="keystone-api" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.672133 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="44f9c92a-dd8e-43e6-a9af-9235ff6e210f" containerName="keystone-api" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.672524 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.683043 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystoned058-account-delete-kl8ff"] Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.760721 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/691d3637-8015-4113-9e46-9ea460b6c18c-operator-scripts\") pod \"keystoned058-account-delete-kl8ff\" (UID: \"691d3637-8015-4113-9e46-9ea460b6c18c\") " pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.760786 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hgzc\" (UniqueName: \"kubernetes.io/projected/691d3637-8015-4113-9e46-9ea460b6c18c-kube-api-access-4hgzc\") pod \"keystoned058-account-delete-kl8ff\" (UID: \"691d3637-8015-4113-9e46-9ea460b6c18c\") " pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.862789 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/691d3637-8015-4113-9e46-9ea460b6c18c-operator-scripts\") pod \"keystoned058-account-delete-kl8ff\" (UID: \"691d3637-8015-4113-9e46-9ea460b6c18c\") " pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.862855 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hgzc\" (UniqueName: \"kubernetes.io/projected/691d3637-8015-4113-9e46-9ea460b6c18c-kube-api-access-4hgzc\") pod \"keystoned058-account-delete-kl8ff\" (UID: \"691d3637-8015-4113-9e46-9ea460b6c18c\") " pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.864492 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/691d3637-8015-4113-9e46-9ea460b6c18c-operator-scripts\") pod \"keystoned058-account-delete-kl8ff\" (UID: \"691d3637-8015-4113-9e46-9ea460b6c18c\") " pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.880932 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hgzc\" (UniqueName: \"kubernetes.io/projected/691d3637-8015-4113-9e46-9ea460b6c18c-kube-api-access-4hgzc\") pod \"keystoned058-account-delete-kl8ff\" (UID: \"691d3637-8015-4113-9e46-9ea460b6c18c\") " pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:21 crc kubenswrapper[4718]: I1206 04:27:21.989056 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:22 crc kubenswrapper[4718]: I1206 04:27:22.401500 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystoned058-account-delete-kl8ff"] Dec 06 04:27:23 crc kubenswrapper[4718]: I1206 04:27:23.045601 4718 generic.go:334] "Generic (PLEG): container finished" podID="691d3637-8015-4113-9e46-9ea460b6c18c" containerID="fbb003bd23b84746ad4abcfe0224e9c2e1c994221ebe6aab3271ece17e9b322f" exitCode=0 Dec 06 04:27:23 crc kubenswrapper[4718]: I1206 04:27:23.045648 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" event={"ID":"691d3637-8015-4113-9e46-9ea460b6c18c","Type":"ContainerDied","Data":"fbb003bd23b84746ad4abcfe0224e9c2e1c994221ebe6aab3271ece17e9b322f"} Dec 06 04:27:23 crc kubenswrapper[4718]: I1206 04:27:23.045699 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" event={"ID":"691d3637-8015-4113-9e46-9ea460b6c18c","Type":"ContainerStarted","Data":"d62746c73bb5a91543726c33b75ca5df2c7716a5ef336c5c71370add8161924f"} Dec 06 04:27:23 crc kubenswrapper[4718]: I1206 04:27:23.335689 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f0aadfe-a9c3-4c89-be50-e91df227c3a2" path="/var/lib/kubelet/pods/2f0aadfe-a9c3-4c89-be50-e91df227c3a2/volumes" Dec 06 04:27:23 crc kubenswrapper[4718]: I1206 04:27:23.336156 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4df8fbb0-37c3-4acb-99fe-1d07364af0b9" path="/var/lib/kubelet/pods/4df8fbb0-37c3-4acb-99fe-1d07364af0b9/volumes" Dec 06 04:27:23 crc kubenswrapper[4718]: I1206 04:27:23.336624 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9" path="/var/lib/kubelet/pods/b7fdbe31-eb3b-4660-b6e2-a8a65e074cb9/volumes" Dec 06 04:27:24 crc kubenswrapper[4718]: I1206 04:27:24.297250 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:24 crc kubenswrapper[4718]: I1206 04:27:24.410096 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/691d3637-8015-4113-9e46-9ea460b6c18c-operator-scripts\") pod \"691d3637-8015-4113-9e46-9ea460b6c18c\" (UID: \"691d3637-8015-4113-9e46-9ea460b6c18c\") " Dec 06 04:27:24 crc kubenswrapper[4718]: I1206 04:27:24.410287 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hgzc\" (UniqueName: \"kubernetes.io/projected/691d3637-8015-4113-9e46-9ea460b6c18c-kube-api-access-4hgzc\") pod \"691d3637-8015-4113-9e46-9ea460b6c18c\" (UID: \"691d3637-8015-4113-9e46-9ea460b6c18c\") " Dec 06 04:27:24 crc kubenswrapper[4718]: I1206 04:27:24.410913 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/691d3637-8015-4113-9e46-9ea460b6c18c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "691d3637-8015-4113-9e46-9ea460b6c18c" (UID: "691d3637-8015-4113-9e46-9ea460b6c18c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:27:24 crc kubenswrapper[4718]: I1206 04:27:24.416304 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/691d3637-8015-4113-9e46-9ea460b6c18c-kube-api-access-4hgzc" (OuterVolumeSpecName: "kube-api-access-4hgzc") pod "691d3637-8015-4113-9e46-9ea460b6c18c" (UID: "691d3637-8015-4113-9e46-9ea460b6c18c"). InnerVolumeSpecName "kube-api-access-4hgzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:27:24 crc kubenswrapper[4718]: I1206 04:27:24.512810 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hgzc\" (UniqueName: \"kubernetes.io/projected/691d3637-8015-4113-9e46-9ea460b6c18c-kube-api-access-4hgzc\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:24 crc kubenswrapper[4718]: I1206 04:27:24.512857 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/691d3637-8015-4113-9e46-9ea460b6c18c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:25 crc kubenswrapper[4718]: I1206 04:27:25.061489 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" event={"ID":"691d3637-8015-4113-9e46-9ea460b6c18c","Type":"ContainerDied","Data":"d62746c73bb5a91543726c33b75ca5df2c7716a5ef336c5c71370add8161924f"} Dec 06 04:27:25 crc kubenswrapper[4718]: I1206 04:27:25.061531 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d62746c73bb5a91543726c33b75ca5df2c7716a5ef336c5c71370add8161924f" Dec 06 04:27:25 crc kubenswrapper[4718]: I1206 04:27:25.061564 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystoned058-account-delete-kl8ff" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.690925 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-zdjw8"] Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.695850 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-zdjw8"] Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.717336 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystoned058-account-delete-kl8ff"] Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.721946 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2"] Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.726202 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-d058-account-create-update-m8cd2"] Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.732430 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystoned058-account-delete-kl8ff"] Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.787406 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ppzf5"] Dec 06 04:27:26 crc kubenswrapper[4718]: E1206 04:27:26.787676 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691d3637-8015-4113-9e46-9ea460b6c18c" containerName="mariadb-account-delete" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.787689 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="691d3637-8015-4113-9e46-9ea460b6c18c" containerName="mariadb-account-delete" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.787811 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="691d3637-8015-4113-9e46-9ea460b6c18c" containerName="mariadb-account-delete" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.788215 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.817858 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ppzf5"] Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.841580 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lp2j\" (UniqueName: \"kubernetes.io/projected/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-kube-api-access-5lp2j\") pod \"keystone-db-create-ppzf5\" (UID: \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\") " pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.841757 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-operator-scripts\") pod \"keystone-db-create-ppzf5\" (UID: \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\") " pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.889976 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx"] Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.891255 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.892823 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.896145 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx"] Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.942731 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-operator-scripts\") pod \"keystone-db-create-ppzf5\" (UID: \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\") " pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.942949 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lp2j\" (UniqueName: \"kubernetes.io/projected/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-kube-api-access-5lp2j\") pod \"keystone-db-create-ppzf5\" (UID: \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\") " pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.943727 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-operator-scripts\") pod \"keystone-db-create-ppzf5\" (UID: \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\") " pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:26 crc kubenswrapper[4718]: I1206 04:27:26.963059 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lp2j\" (UniqueName: \"kubernetes.io/projected/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-kube-api-access-5lp2j\") pod \"keystone-db-create-ppzf5\" (UID: \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\") " pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.044772 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95197796-ac9e-47d1-b19f-92b534a42e5a-operator-scripts\") pod \"keystone-178c-account-create-update-8tkqx\" (UID: \"95197796-ac9e-47d1-b19f-92b534a42e5a\") " pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.044881 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2tdh\" (UniqueName: \"kubernetes.io/projected/95197796-ac9e-47d1-b19f-92b534a42e5a-kube-api-access-x2tdh\") pod \"keystone-178c-account-create-update-8tkqx\" (UID: \"95197796-ac9e-47d1-b19f-92b534a42e5a\") " pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.114488 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.145742 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95197796-ac9e-47d1-b19f-92b534a42e5a-operator-scripts\") pod \"keystone-178c-account-create-update-8tkqx\" (UID: \"95197796-ac9e-47d1-b19f-92b534a42e5a\") " pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.145884 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2tdh\" (UniqueName: \"kubernetes.io/projected/95197796-ac9e-47d1-b19f-92b534a42e5a-kube-api-access-x2tdh\") pod \"keystone-178c-account-create-update-8tkqx\" (UID: \"95197796-ac9e-47d1-b19f-92b534a42e5a\") " pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.146504 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95197796-ac9e-47d1-b19f-92b534a42e5a-operator-scripts\") pod \"keystone-178c-account-create-update-8tkqx\" (UID: \"95197796-ac9e-47d1-b19f-92b534a42e5a\") " pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.164057 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2tdh\" (UniqueName: \"kubernetes.io/projected/95197796-ac9e-47d1-b19f-92b534a42e5a-kube-api-access-x2tdh\") pod \"keystone-178c-account-create-update-8tkqx\" (UID: \"95197796-ac9e-47d1-b19f-92b534a42e5a\") " pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.213342 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.338075 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="183c06e8-ca44-4a1e-b6c9-c7599077370a" path="/var/lib/kubelet/pods/183c06e8-ca44-4a1e-b6c9-c7599077370a/volumes" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.338859 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="691d3637-8015-4113-9e46-9ea460b6c18c" path="/var/lib/kubelet/pods/691d3637-8015-4113-9e46-9ea460b6c18c/volumes" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.339499 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb6639e4-b470-45ca-8717-3f231badd399" path="/var/lib/kubelet/pods/fb6639e4-b470-45ca-8717-3f231badd399/volumes" Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.580808 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ppzf5"] Dec 06 04:27:27 crc kubenswrapper[4718]: W1206 04:27:27.660772 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95197796_ac9e_47d1_b19f_92b534a42e5a.slice/crio-7a60e796dda5b674abf46d8172cf08b45312748ec0e7b2b7f92ff44511fa0764 WatchSource:0}: Error finding container 7a60e796dda5b674abf46d8172cf08b45312748ec0e7b2b7f92ff44511fa0764: Status 404 returned error can't find the container with id 7a60e796dda5b674abf46d8172cf08b45312748ec0e7b2b7f92ff44511fa0764 Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.660970 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx"] Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.876942 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:27:27 crc kubenswrapper[4718]: I1206 04:27:27.876996 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:27:28 crc kubenswrapper[4718]: I1206 04:27:28.084717 4718 generic.go:334] "Generic (PLEG): container finished" podID="95197796-ac9e-47d1-b19f-92b534a42e5a" containerID="b0099011249fdf367d16392d4329508d27b73d1cef02576982f565954ecb143b" exitCode=0 Dec 06 04:27:28 crc kubenswrapper[4718]: I1206 04:27:28.084808 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" event={"ID":"95197796-ac9e-47d1-b19f-92b534a42e5a","Type":"ContainerDied","Data":"b0099011249fdf367d16392d4329508d27b73d1cef02576982f565954ecb143b"} Dec 06 04:27:28 crc kubenswrapper[4718]: I1206 04:27:28.084850 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" event={"ID":"95197796-ac9e-47d1-b19f-92b534a42e5a","Type":"ContainerStarted","Data":"7a60e796dda5b674abf46d8172cf08b45312748ec0e7b2b7f92ff44511fa0764"} Dec 06 04:27:28 crc kubenswrapper[4718]: I1206 04:27:28.086602 4718 generic.go:334] "Generic (PLEG): container finished" podID="81fb977d-4fcb-4fde-9719-7d6d2b9dfae0" containerID="9f02663b460b91cef7ba90f1d492a0f28e806adc30369a70208dcb0f584f4390" exitCode=0 Dec 06 04:27:28 crc kubenswrapper[4718]: I1206 04:27:28.086667 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-ppzf5" event={"ID":"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0","Type":"ContainerDied","Data":"9f02663b460b91cef7ba90f1d492a0f28e806adc30369a70208dcb0f584f4390"} Dec 06 04:27:28 crc kubenswrapper[4718]: I1206 04:27:28.086877 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-ppzf5" event={"ID":"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0","Type":"ContainerStarted","Data":"c81f3e902fccf2d4e001b2a2a337b74b6e24f8a0f00e8185e45f094058b1fd1e"} Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.414511 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.418334 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.575465 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95197796-ac9e-47d1-b19f-92b534a42e5a-operator-scripts\") pod \"95197796-ac9e-47d1-b19f-92b534a42e5a\" (UID: \"95197796-ac9e-47d1-b19f-92b534a42e5a\") " Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.575717 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lp2j\" (UniqueName: \"kubernetes.io/projected/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-kube-api-access-5lp2j\") pod \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\" (UID: \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\") " Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.575795 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-operator-scripts\") pod \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\" (UID: \"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0\") " Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.575834 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2tdh\" (UniqueName: \"kubernetes.io/projected/95197796-ac9e-47d1-b19f-92b534a42e5a-kube-api-access-x2tdh\") pod \"95197796-ac9e-47d1-b19f-92b534a42e5a\" (UID: \"95197796-ac9e-47d1-b19f-92b534a42e5a\") " Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.576169 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95197796-ac9e-47d1-b19f-92b534a42e5a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95197796-ac9e-47d1-b19f-92b534a42e5a" (UID: "95197796-ac9e-47d1-b19f-92b534a42e5a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.576568 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81fb977d-4fcb-4fde-9719-7d6d2b9dfae0" (UID: "81fb977d-4fcb-4fde-9719-7d6d2b9dfae0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.580754 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-kube-api-access-5lp2j" (OuterVolumeSpecName: "kube-api-access-5lp2j") pod "81fb977d-4fcb-4fde-9719-7d6d2b9dfae0" (UID: "81fb977d-4fcb-4fde-9719-7d6d2b9dfae0"). InnerVolumeSpecName "kube-api-access-5lp2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.582201 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95197796-ac9e-47d1-b19f-92b534a42e5a-kube-api-access-x2tdh" (OuterVolumeSpecName: "kube-api-access-x2tdh") pod "95197796-ac9e-47d1-b19f-92b534a42e5a" (UID: "95197796-ac9e-47d1-b19f-92b534a42e5a"). InnerVolumeSpecName "kube-api-access-x2tdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.677381 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lp2j\" (UniqueName: \"kubernetes.io/projected/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-kube-api-access-5lp2j\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.677420 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.677434 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2tdh\" (UniqueName: \"kubernetes.io/projected/95197796-ac9e-47d1-b19f-92b534a42e5a-kube-api-access-x2tdh\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:29 crc kubenswrapper[4718]: I1206 04:27:29.677446 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95197796-ac9e-47d1-b19f-92b534a42e5a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:30 crc kubenswrapper[4718]: I1206 04:27:30.124855 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" Dec 06 04:27:30 crc kubenswrapper[4718]: I1206 04:27:30.124884 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx" event={"ID":"95197796-ac9e-47d1-b19f-92b534a42e5a","Type":"ContainerDied","Data":"7a60e796dda5b674abf46d8172cf08b45312748ec0e7b2b7f92ff44511fa0764"} Dec 06 04:27:30 crc kubenswrapper[4718]: I1206 04:27:30.124969 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a60e796dda5b674abf46d8172cf08b45312748ec0e7b2b7f92ff44511fa0764" Dec 06 04:27:30 crc kubenswrapper[4718]: I1206 04:27:30.128419 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-ppzf5" event={"ID":"81fb977d-4fcb-4fde-9719-7d6d2b9dfae0","Type":"ContainerDied","Data":"c81f3e902fccf2d4e001b2a2a337b74b6e24f8a0f00e8185e45f094058b1fd1e"} Dec 06 04:27:30 crc kubenswrapper[4718]: I1206 04:27:30.128478 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c81f3e902fccf2d4e001b2a2a337b74b6e24f8a0f00e8185e45f094058b1fd1e" Dec 06 04:27:30 crc kubenswrapper[4718]: I1206 04:27:30.128573 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-ppzf5" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.443533 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-jffrv"] Dec 06 04:27:32 crc kubenswrapper[4718]: E1206 04:27:32.444225 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95197796-ac9e-47d1-b19f-92b534a42e5a" containerName="mariadb-account-create-update" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.444284 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="95197796-ac9e-47d1-b19f-92b534a42e5a" containerName="mariadb-account-create-update" Dec 06 04:27:32 crc kubenswrapper[4718]: E1206 04:27:32.444321 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fb977d-4fcb-4fde-9719-7d6d2b9dfae0" containerName="mariadb-database-create" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.444338 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fb977d-4fcb-4fde-9719-7d6d2b9dfae0" containerName="mariadb-database-create" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.444554 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="95197796-ac9e-47d1-b19f-92b534a42e5a" containerName="mariadb-account-create-update" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.444583 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="81fb977d-4fcb-4fde-9719-7d6d2b9dfae0" containerName="mariadb-database-create" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.446988 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.449815 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"combined-ca-bundle" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.450140 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.450561 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.450797 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-7d5tv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.454652 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.465823 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-jffrv"] Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.519302 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-config-data\") pod \"keystone-db-sync-jffrv\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.519404 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-combined-ca-bundle\") pod \"keystone-db-sync-jffrv\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.519448 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m2kx\" (UniqueName: \"kubernetes.io/projected/cf96ccf4-fa66-481e-a910-8e64d723def9-kube-api-access-2m2kx\") pod \"keystone-db-sync-jffrv\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.621573 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-config-data\") pod \"keystone-db-sync-jffrv\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.621650 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-combined-ca-bundle\") pod \"keystone-db-sync-jffrv\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.621681 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m2kx\" (UniqueName: \"kubernetes.io/projected/cf96ccf4-fa66-481e-a910-8e64d723def9-kube-api-access-2m2kx\") pod \"keystone-db-sync-jffrv\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.631222 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-combined-ca-bundle\") pod \"keystone-db-sync-jffrv\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.633420 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-config-data\") pod \"keystone-db-sync-jffrv\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.637940 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m2kx\" (UniqueName: \"kubernetes.io/projected/cf96ccf4-fa66-481e-a910-8e64d723def9-kube-api-access-2m2kx\") pod \"keystone-db-sync-jffrv\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:32 crc kubenswrapper[4718]: I1206 04:27:32.772965 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:33 crc kubenswrapper[4718]: I1206 04:27:33.190980 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-jffrv"] Dec 06 04:27:34 crc kubenswrapper[4718]: I1206 04:27:34.161638 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-jffrv" event={"ID":"cf96ccf4-fa66-481e-a910-8e64d723def9","Type":"ContainerStarted","Data":"fe35bcbf986528a9205830bab0a4d15aea064afa048b10900cdf0cf8ad6c9ae5"} Dec 06 04:27:36 crc kubenswrapper[4718]: I1206 04:27:36.182371 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-jffrv" event={"ID":"cf96ccf4-fa66-481e-a910-8e64d723def9","Type":"ContainerStarted","Data":"c1c5e05f6a64a1722daadfe2dfc587e9395af4b532db00f918ba84d596e9498a"} Dec 06 04:27:36 crc kubenswrapper[4718]: I1206 04:27:36.213041 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-db-sync-jffrv" podStartSLOduration=4.213020417 podStartE2EDuration="4.213020417s" podCreationTimestamp="2025-12-06 04:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:27:36.205955974 +0000 UTC m=+1245.211661145" watchObservedRunningTime="2025-12-06 04:27:36.213020417 +0000 UTC m=+1245.218725588" Dec 06 04:27:38 crc kubenswrapper[4718]: I1206 04:27:38.206052 4718 generic.go:334] "Generic (PLEG): container finished" podID="cf96ccf4-fa66-481e-a910-8e64d723def9" containerID="c1c5e05f6a64a1722daadfe2dfc587e9395af4b532db00f918ba84d596e9498a" exitCode=0 Dec 06 04:27:38 crc kubenswrapper[4718]: I1206 04:27:38.206164 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-jffrv" event={"ID":"cf96ccf4-fa66-481e-a910-8e64d723def9","Type":"ContainerDied","Data":"c1c5e05f6a64a1722daadfe2dfc587e9395af4b532db00f918ba84d596e9498a"} Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.492817 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.634926 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m2kx\" (UniqueName: \"kubernetes.io/projected/cf96ccf4-fa66-481e-a910-8e64d723def9-kube-api-access-2m2kx\") pod \"cf96ccf4-fa66-481e-a910-8e64d723def9\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.635187 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-combined-ca-bundle\") pod \"cf96ccf4-fa66-481e-a910-8e64d723def9\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.635295 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-config-data\") pod \"cf96ccf4-fa66-481e-a910-8e64d723def9\" (UID: \"cf96ccf4-fa66-481e-a910-8e64d723def9\") " Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.643456 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf96ccf4-fa66-481e-a910-8e64d723def9-kube-api-access-2m2kx" (OuterVolumeSpecName: "kube-api-access-2m2kx") pod "cf96ccf4-fa66-481e-a910-8e64d723def9" (UID: "cf96ccf4-fa66-481e-a910-8e64d723def9"). InnerVolumeSpecName "kube-api-access-2m2kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.677561 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf96ccf4-fa66-481e-a910-8e64d723def9" (UID: "cf96ccf4-fa66-481e-a910-8e64d723def9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.695440 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-config-data" (OuterVolumeSpecName: "config-data") pod "cf96ccf4-fa66-481e-a910-8e64d723def9" (UID: "cf96ccf4-fa66-481e-a910-8e64d723def9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.737590 4718 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.737639 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf96ccf4-fa66-481e-a910-8e64d723def9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:39 crc kubenswrapper[4718]: I1206 04:27:39.737661 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m2kx\" (UniqueName: \"kubernetes.io/projected/cf96ccf4-fa66-481e-a910-8e64d723def9-kube-api-access-2m2kx\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.228121 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-jffrv" event={"ID":"cf96ccf4-fa66-481e-a910-8e64d723def9","Type":"ContainerDied","Data":"fe35bcbf986528a9205830bab0a4d15aea064afa048b10900cdf0cf8ad6c9ae5"} Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.228502 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe35bcbf986528a9205830bab0a4d15aea064afa048b10900cdf0cf8ad6c9ae5" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.228293 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-jffrv" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.581207 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-tn8wz"] Dec 06 04:27:40 crc kubenswrapper[4718]: E1206 04:27:40.582174 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf96ccf4-fa66-481e-a910-8e64d723def9" containerName="keystone-db-sync" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.582204 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf96ccf4-fa66-481e-a910-8e64d723def9" containerName="keystone-db-sync" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.582504 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf96ccf4-fa66-481e-a910-8e64d723def9" containerName="keystone-db-sync" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.583416 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.589018 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.589130 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-7d5tv" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.589307 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.589449 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"combined-ca-bundle" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.589576 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.589592 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.602456 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-tn8wz"] Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.752490 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6bdp\" (UniqueName: \"kubernetes.io/projected/482282f2-2ad3-487c-b756-5a2e874288ee-kube-api-access-p6bdp\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.752546 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-scripts\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.752584 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-combined-ca-bundle\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.752738 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-fernet-keys\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.752924 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-config-data\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.753002 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-credential-keys\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.854629 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-credential-keys\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.854695 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6bdp\" (UniqueName: \"kubernetes.io/projected/482282f2-2ad3-487c-b756-5a2e874288ee-kube-api-access-p6bdp\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.854739 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-scripts\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.854807 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-combined-ca-bundle\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.854858 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-fernet-keys\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.854943 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-config-data\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.859637 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-scripts\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.860526 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-config-data\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.866872 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-fernet-keys\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.867026 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-credential-keys\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.867170 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-combined-ca-bundle\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.873791 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6bdp\" (UniqueName: \"kubernetes.io/projected/482282f2-2ad3-487c-b756-5a2e874288ee-kube-api-access-p6bdp\") pod \"keystone-bootstrap-tn8wz\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:40 crc kubenswrapper[4718]: I1206 04:27:40.908763 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:41 crc kubenswrapper[4718]: I1206 04:27:41.353194 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-tn8wz"] Dec 06 04:27:41 crc kubenswrapper[4718]: W1206 04:27:41.365844 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod482282f2_2ad3_487c_b756_5a2e874288ee.slice/crio-808d6b371726cc31d8ab55ff6cafcb997c807f77e1685597ce185e0d74a704fd WatchSource:0}: Error finding container 808d6b371726cc31d8ab55ff6cafcb997c807f77e1685597ce185e0d74a704fd: Status 404 returned error can't find the container with id 808d6b371726cc31d8ab55ff6cafcb997c807f77e1685597ce185e0d74a704fd Dec 06 04:27:42 crc kubenswrapper[4718]: I1206 04:27:42.244820 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" event={"ID":"482282f2-2ad3-487c-b756-5a2e874288ee","Type":"ContainerStarted","Data":"275894c18f169adb9c38bb81d34a65418eab583a1a05a8154ce5d60c2ff28dae"} Dec 06 04:27:42 crc kubenswrapper[4718]: I1206 04:27:42.245200 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" event={"ID":"482282f2-2ad3-487c-b756-5a2e874288ee","Type":"ContainerStarted","Data":"808d6b371726cc31d8ab55ff6cafcb997c807f77e1685597ce185e0d74a704fd"} Dec 06 04:27:42 crc kubenswrapper[4718]: I1206 04:27:42.264632 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" podStartSLOduration=2.264608817 podStartE2EDuration="2.264608817s" podCreationTimestamp="2025-12-06 04:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:27:42.26036357 +0000 UTC m=+1251.266068731" watchObservedRunningTime="2025-12-06 04:27:42.264608817 +0000 UTC m=+1251.270314008" Dec 06 04:27:45 crc kubenswrapper[4718]: I1206 04:27:45.275714 4718 generic.go:334] "Generic (PLEG): container finished" podID="482282f2-2ad3-487c-b756-5a2e874288ee" containerID="275894c18f169adb9c38bb81d34a65418eab583a1a05a8154ce5d60c2ff28dae" exitCode=0 Dec 06 04:27:45 crc kubenswrapper[4718]: I1206 04:27:45.275833 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" event={"ID":"482282f2-2ad3-487c-b756-5a2e874288ee","Type":"ContainerDied","Data":"275894c18f169adb9c38bb81d34a65418eab583a1a05a8154ce5d60c2ff28dae"} Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.588524 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.748171 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-credential-keys\") pod \"482282f2-2ad3-487c-b756-5a2e874288ee\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.748321 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-config-data\") pod \"482282f2-2ad3-487c-b756-5a2e874288ee\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.748342 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-scripts\") pod \"482282f2-2ad3-487c-b756-5a2e874288ee\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.748366 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6bdp\" (UniqueName: \"kubernetes.io/projected/482282f2-2ad3-487c-b756-5a2e874288ee-kube-api-access-p6bdp\") pod \"482282f2-2ad3-487c-b756-5a2e874288ee\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.748397 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-combined-ca-bundle\") pod \"482282f2-2ad3-487c-b756-5a2e874288ee\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.748424 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-fernet-keys\") pod \"482282f2-2ad3-487c-b756-5a2e874288ee\" (UID: \"482282f2-2ad3-487c-b756-5a2e874288ee\") " Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.755382 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-scripts" (OuterVolumeSpecName: "scripts") pod "482282f2-2ad3-487c-b756-5a2e874288ee" (UID: "482282f2-2ad3-487c-b756-5a2e874288ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.755411 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "482282f2-2ad3-487c-b756-5a2e874288ee" (UID: "482282f2-2ad3-487c-b756-5a2e874288ee"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.755435 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "482282f2-2ad3-487c-b756-5a2e874288ee" (UID: "482282f2-2ad3-487c-b756-5a2e874288ee"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.755626 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/482282f2-2ad3-487c-b756-5a2e874288ee-kube-api-access-p6bdp" (OuterVolumeSpecName: "kube-api-access-p6bdp") pod "482282f2-2ad3-487c-b756-5a2e874288ee" (UID: "482282f2-2ad3-487c-b756-5a2e874288ee"). InnerVolumeSpecName "kube-api-access-p6bdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.768573 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "482282f2-2ad3-487c-b756-5a2e874288ee" (UID: "482282f2-2ad3-487c-b756-5a2e874288ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.771649 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-config-data" (OuterVolumeSpecName: "config-data") pod "482282f2-2ad3-487c-b756-5a2e874288ee" (UID: "482282f2-2ad3-487c-b756-5a2e874288ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.849607 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.849639 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.849648 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.849657 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6bdp\" (UniqueName: \"kubernetes.io/projected/482282f2-2ad3-487c-b756-5a2e874288ee-kube-api-access-p6bdp\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.849668 4718 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:46 crc kubenswrapper[4718]: I1206 04:27:46.849675 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/482282f2-2ad3-487c-b756-5a2e874288ee-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.292540 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" event={"ID":"482282f2-2ad3-487c-b756-5a2e874288ee","Type":"ContainerDied","Data":"808d6b371726cc31d8ab55ff6cafcb997c807f77e1685597ce185e0d74a704fd"} Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.292778 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="808d6b371726cc31d8ab55ff6cafcb997c807f77e1685597ce185e0d74a704fd" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.292640 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-tn8wz" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.378338 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-5c775cf77-f96nh"] Dec 06 04:27:47 crc kubenswrapper[4718]: E1206 04:27:47.378630 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482282f2-2ad3-487c-b756-5a2e874288ee" containerName="keystone-bootstrap" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.378652 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="482282f2-2ad3-487c-b756-5a2e874288ee" containerName="keystone-bootstrap" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.378812 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="482282f2-2ad3-487c-b756-5a2e874288ee" containerName="keystone-bootstrap" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.379389 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.381145 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.381182 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.381971 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"cert-keystone-internal-svc" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.382166 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"combined-ca-bundle" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.382340 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"cert-keystone-public-svc" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.382372 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.383828 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-7d5tv" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.392273 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-5c775cf77-f96nh"] Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.466256 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-fernet-keys\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.466311 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7xwh\" (UniqueName: \"kubernetes.io/projected/14645e66-70f7-411f-89cb-1cfabf743478-kube-api-access-v7xwh\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.466344 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-combined-ca-bundle\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.466373 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-scripts\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.466419 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-config-data\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.466435 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-credential-keys\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.466450 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-public-tls-certs\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.466471 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-internal-tls-certs\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.567487 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-fernet-keys\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.567549 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7xwh\" (UniqueName: \"kubernetes.io/projected/14645e66-70f7-411f-89cb-1cfabf743478-kube-api-access-v7xwh\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.567590 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-combined-ca-bundle\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.567620 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-scripts\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.567657 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-config-data\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.567676 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-credential-keys\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.567697 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-public-tls-certs\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.567731 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-internal-tls-certs\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.571946 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-combined-ca-bundle\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.572346 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-config-data\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.572986 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-public-tls-certs\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.572914 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-fernet-keys\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.573274 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-scripts\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.574974 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-credential-keys\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.576657 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-internal-tls-certs\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.585864 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7xwh\" (UniqueName: \"kubernetes.io/projected/14645e66-70f7-411f-89cb-1cfabf743478-kube-api-access-v7xwh\") pod \"keystone-5c775cf77-f96nh\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:47 crc kubenswrapper[4718]: I1206 04:27:47.696095 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:48 crc kubenswrapper[4718]: I1206 04:27:48.117301 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-5c775cf77-f96nh"] Dec 06 04:27:48 crc kubenswrapper[4718]: I1206 04:27:48.299464 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" event={"ID":"14645e66-70f7-411f-89cb-1cfabf743478","Type":"ContainerStarted","Data":"69b63161ca5434b57b252b9e803be3700239f51d7e8c3deaf44f3bf8b5ffc114"} Dec 06 04:27:49 crc kubenswrapper[4718]: I1206 04:27:49.306811 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" event={"ID":"14645e66-70f7-411f-89cb-1cfabf743478","Type":"ContainerStarted","Data":"c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4"} Dec 06 04:27:49 crc kubenswrapper[4718]: I1206 04:27:49.306953 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:27:49 crc kubenswrapper[4718]: I1206 04:27:49.332021 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" podStartSLOduration=2.332001876 podStartE2EDuration="2.332001876s" podCreationTimestamp="2025-12-06 04:27:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:27:49.322492816 +0000 UTC m=+1258.328197977" watchObservedRunningTime="2025-12-06 04:27:49.332001876 +0000 UTC m=+1258.337707037" Dec 06 04:27:57 crc kubenswrapper[4718]: I1206 04:27:57.877307 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:27:57 crc kubenswrapper[4718]: I1206 04:27:57.877932 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:27:57 crc kubenswrapper[4718]: I1206 04:27:57.877998 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:27:57 crc kubenswrapper[4718]: I1206 04:27:57.878856 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d135c41c1c0b733890b2561029cffaa9956d80b356ad7dc0217957931de3b87"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:27:57 crc kubenswrapper[4718]: I1206 04:27:57.878947 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://8d135c41c1c0b733890b2561029cffaa9956d80b356ad7dc0217957931de3b87" gracePeriod=600 Dec 06 04:27:58 crc kubenswrapper[4718]: I1206 04:27:58.381607 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="8d135c41c1c0b733890b2561029cffaa9956d80b356ad7dc0217957931de3b87" exitCode=0 Dec 06 04:27:58 crc kubenswrapper[4718]: I1206 04:27:58.381874 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"8d135c41c1c0b733890b2561029cffaa9956d80b356ad7dc0217957931de3b87"} Dec 06 04:27:58 crc kubenswrapper[4718]: I1206 04:27:58.382072 4718 scope.go:117] "RemoveContainer" containerID="5492a0b85bf17b4550c5ff3ce07a61828c2b6f5eb097631d3cdfc69dc9a69c85" Dec 06 04:27:59 crc kubenswrapper[4718]: I1206 04:27:59.392849 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"d81f022389d14fa68d0d6add32cf0d12240d0d061eea1e86215bd5dbdd1355c2"} Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.132863 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.476146 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-tn8wz"] Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.484455 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-tn8wz"] Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.491375 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-jffrv"] Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.499730 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-jffrv"] Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.506368 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-5c775cf77-f96nh"] Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.538732 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone178c-account-delete-wm4rd"] Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.539422 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.554613 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone178c-account-delete-wm4rd"] Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.589755 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" podUID="14645e66-70f7-411f-89cb-1cfabf743478" containerName="keystone-api" containerID="cri-o://c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4" gracePeriod=30 Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.617128 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6r8v\" (UniqueName: \"kubernetes.io/projected/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-kube-api-access-m6r8v\") pod \"keystone178c-account-delete-wm4rd\" (UID: \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\") " pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.617265 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-operator-scripts\") pod \"keystone178c-account-delete-wm4rd\" (UID: \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\") " pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.718077 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6r8v\" (UniqueName: \"kubernetes.io/projected/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-kube-api-access-m6r8v\") pod \"keystone178c-account-delete-wm4rd\" (UID: \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\") " pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.718209 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-operator-scripts\") pod \"keystone178c-account-delete-wm4rd\" (UID: \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\") " pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.718872 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-operator-scripts\") pod \"keystone178c-account-delete-wm4rd\" (UID: \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\") " pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.742554 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6r8v\" (UniqueName: \"kubernetes.io/projected/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-kube-api-access-m6r8v\") pod \"keystone178c-account-delete-wm4rd\" (UID: \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\") " pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:19 crc kubenswrapper[4718]: I1206 04:28:19.862696 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:20 crc kubenswrapper[4718]: I1206 04:28:20.323407 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone178c-account-delete-wm4rd"] Dec 06 04:28:20 crc kubenswrapper[4718]: I1206 04:28:20.600729 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" event={"ID":"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87","Type":"ContainerStarted","Data":"8692c0d26ab0ffc9e45c3d9d400c3b3b032bfa2cdd828a9e670dfa5810e3bdc2"} Dec 06 04:28:20 crc kubenswrapper[4718]: I1206 04:28:20.601150 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" event={"ID":"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87","Type":"ContainerStarted","Data":"85cb157ff654591d8a29271a3da9675064fbaf6d5f7a6d2ad98ae0c9794133f1"} Dec 06 04:28:20 crc kubenswrapper[4718]: I1206 04:28:20.626102 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" podStartSLOduration=1.626081162 podStartE2EDuration="1.626081162s" podCreationTimestamp="2025-12-06 04:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:28:20.61870968 +0000 UTC m=+1289.624414871" watchObservedRunningTime="2025-12-06 04:28:20.626081162 +0000 UTC m=+1289.631786323" Dec 06 04:28:21 crc kubenswrapper[4718]: I1206 04:28:21.340012 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="482282f2-2ad3-487c-b756-5a2e874288ee" path="/var/lib/kubelet/pods/482282f2-2ad3-487c-b756-5a2e874288ee/volumes" Dec 06 04:28:21 crc kubenswrapper[4718]: I1206 04:28:21.340948 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf96ccf4-fa66-481e-a910-8e64d723def9" path="/var/lib/kubelet/pods/cf96ccf4-fa66-481e-a910-8e64d723def9/volumes" Dec 06 04:28:21 crc kubenswrapper[4718]: I1206 04:28:21.609178 4718 generic.go:334] "Generic (PLEG): container finished" podID="124214bc-17e4-4e8b-9c4b-1f52e0ccbd87" containerID="8692c0d26ab0ffc9e45c3d9d400c3b3b032bfa2cdd828a9e670dfa5810e3bdc2" exitCode=0 Dec 06 04:28:21 crc kubenswrapper[4718]: I1206 04:28:21.609252 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" event={"ID":"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87","Type":"ContainerDied","Data":"8692c0d26ab0ffc9e45c3d9d400c3b3b032bfa2cdd828a9e670dfa5810e3bdc2"} Dec 06 04:28:22 crc kubenswrapper[4718]: I1206 04:28:22.958864 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.067604 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-operator-scripts\") pod \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\" (UID: \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.067745 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6r8v\" (UniqueName: \"kubernetes.io/projected/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-kube-api-access-m6r8v\") pod \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\" (UID: \"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.068103 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "124214bc-17e4-4e8b-9c4b-1f52e0ccbd87" (UID: "124214bc-17e4-4e8b-9c4b-1f52e0ccbd87"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.068270 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.072636 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-kube-api-access-m6r8v" (OuterVolumeSpecName: "kube-api-access-m6r8v") pod "124214bc-17e4-4e8b-9c4b-1f52e0ccbd87" (UID: "124214bc-17e4-4e8b-9c4b-1f52e0ccbd87"). InnerVolumeSpecName "kube-api-access-m6r8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.121728 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.169818 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6r8v\" (UniqueName: \"kubernetes.io/projected/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87-kube-api-access-m6r8v\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.271047 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-config-data\") pod \"14645e66-70f7-411f-89cb-1cfabf743478\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.271117 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-credential-keys\") pod \"14645e66-70f7-411f-89cb-1cfabf743478\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.271149 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-fernet-keys\") pod \"14645e66-70f7-411f-89cb-1cfabf743478\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.271670 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7xwh\" (UniqueName: \"kubernetes.io/projected/14645e66-70f7-411f-89cb-1cfabf743478-kube-api-access-v7xwh\") pod \"14645e66-70f7-411f-89cb-1cfabf743478\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.271704 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-scripts\") pod \"14645e66-70f7-411f-89cb-1cfabf743478\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.271724 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-public-tls-certs\") pod \"14645e66-70f7-411f-89cb-1cfabf743478\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.271819 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-combined-ca-bundle\") pod \"14645e66-70f7-411f-89cb-1cfabf743478\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.271874 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-internal-tls-certs\") pod \"14645e66-70f7-411f-89cb-1cfabf743478\" (UID: \"14645e66-70f7-411f-89cb-1cfabf743478\") " Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.274400 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "14645e66-70f7-411f-89cb-1cfabf743478" (UID: "14645e66-70f7-411f-89cb-1cfabf743478"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.274684 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14645e66-70f7-411f-89cb-1cfabf743478-kube-api-access-v7xwh" (OuterVolumeSpecName: "kube-api-access-v7xwh") pod "14645e66-70f7-411f-89cb-1cfabf743478" (UID: "14645e66-70f7-411f-89cb-1cfabf743478"). InnerVolumeSpecName "kube-api-access-v7xwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.275097 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "14645e66-70f7-411f-89cb-1cfabf743478" (UID: "14645e66-70f7-411f-89cb-1cfabf743478"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.276581 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-scripts" (OuterVolumeSpecName: "scripts") pod "14645e66-70f7-411f-89cb-1cfabf743478" (UID: "14645e66-70f7-411f-89cb-1cfabf743478"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.288995 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14645e66-70f7-411f-89cb-1cfabf743478" (UID: "14645e66-70f7-411f-89cb-1cfabf743478"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.291187 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-config-data" (OuterVolumeSpecName: "config-data") pod "14645e66-70f7-411f-89cb-1cfabf743478" (UID: "14645e66-70f7-411f-89cb-1cfabf743478"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.303201 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "14645e66-70f7-411f-89cb-1cfabf743478" (UID: "14645e66-70f7-411f-89cb-1cfabf743478"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.304440 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "14645e66-70f7-411f-89cb-1cfabf743478" (UID: "14645e66-70f7-411f-89cb-1cfabf743478"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.373138 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.373200 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.373219 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.373268 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7xwh\" (UniqueName: \"kubernetes.io/projected/14645e66-70f7-411f-89cb-1cfabf743478-kube-api-access-v7xwh\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.373287 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.373302 4718 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.373319 4718 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.373339 4718 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14645e66-70f7-411f-89cb-1cfabf743478-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.632438 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" event={"ID":"124214bc-17e4-4e8b-9c4b-1f52e0ccbd87","Type":"ContainerDied","Data":"85cb157ff654591d8a29271a3da9675064fbaf6d5f7a6d2ad98ae0c9794133f1"} Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.632473 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone178c-account-delete-wm4rd" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.632499 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85cb157ff654591d8a29271a3da9675064fbaf6d5f7a6d2ad98ae0c9794133f1" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.635102 4718 generic.go:334] "Generic (PLEG): container finished" podID="14645e66-70f7-411f-89cb-1cfabf743478" containerID="c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4" exitCode=0 Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.635164 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" event={"ID":"14645e66-70f7-411f-89cb-1cfabf743478","Type":"ContainerDied","Data":"c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4"} Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.635181 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.635204 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-5c775cf77-f96nh" event={"ID":"14645e66-70f7-411f-89cb-1cfabf743478","Type":"ContainerDied","Data":"69b63161ca5434b57b252b9e803be3700239f51d7e8c3deaf44f3bf8b5ffc114"} Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.635277 4718 scope.go:117] "RemoveContainer" containerID="c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.685523 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-5c775cf77-f96nh"] Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.690200 4718 scope.go:117] "RemoveContainer" containerID="c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4" Dec 06 04:28:23 crc kubenswrapper[4718]: E1206 04:28:23.690619 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4\": container with ID starting with c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4 not found: ID does not exist" containerID="c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.690658 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4"} err="failed to get container status \"c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4\": rpc error: code = NotFound desc = could not find container \"c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4\": container with ID starting with c1e663168fb5f44b1fee82e6b6e0b8c048c98ef24f2374c1961715ec7fc1ebb4 not found: ID does not exist" Dec 06 04:28:23 crc kubenswrapper[4718]: I1206 04:28:23.695017 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-5c775cf77-f96nh"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.596022 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ppzf5"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.606665 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ppzf5"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.614814 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone178c-account-delete-wm4rd"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.620312 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.625512 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone178c-account-delete-wm4rd"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.629841 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-178c-account-create-update-8tkqx"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.837381 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-h9dnk"] Dec 06 04:28:24 crc kubenswrapper[4718]: E1206 04:28:24.837883 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="124214bc-17e4-4e8b-9c4b-1f52e0ccbd87" containerName="mariadb-account-delete" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.837936 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="124214bc-17e4-4e8b-9c4b-1f52e0ccbd87" containerName="mariadb-account-delete" Dec 06 04:28:24 crc kubenswrapper[4718]: E1206 04:28:24.837973 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14645e66-70f7-411f-89cb-1cfabf743478" containerName="keystone-api" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.837988 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="14645e66-70f7-411f-89cb-1cfabf743478" containerName="keystone-api" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.838119 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="124214bc-17e4-4e8b-9c4b-1f52e0ccbd87" containerName="mariadb-account-delete" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.838139 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="14645e66-70f7-411f-89cb-1cfabf743478" containerName="keystone-api" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.838707 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.841675 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.842401 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.880107 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.885843 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-h9dnk"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.892805 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6"] Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.998961 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzp5t\" (UniqueName: \"kubernetes.io/projected/1d319fb8-91c8-4b37-9f5d-11df7656781e-kube-api-access-tzp5t\") pod \"keystone-1a4f-account-create-update-lcjf6\" (UID: \"1d319fb8-91c8-4b37-9f5d-11df7656781e\") " pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.999112 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d319fb8-91c8-4b37-9f5d-11df7656781e-operator-scripts\") pod \"keystone-1a4f-account-create-update-lcjf6\" (UID: \"1d319fb8-91c8-4b37-9f5d-11df7656781e\") " pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.999150 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdqbt\" (UniqueName: \"kubernetes.io/projected/637ba10f-b8e0-48d9-9ea4-96569643bd29-kube-api-access-vdqbt\") pod \"keystone-db-create-h9dnk\" (UID: \"637ba10f-b8e0-48d9-9ea4-96569643bd29\") " pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:24 crc kubenswrapper[4718]: I1206 04:28:24.999199 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/637ba10f-b8e0-48d9-9ea4-96569643bd29-operator-scripts\") pod \"keystone-db-create-h9dnk\" (UID: \"637ba10f-b8e0-48d9-9ea4-96569643bd29\") " pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.100718 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/637ba10f-b8e0-48d9-9ea4-96569643bd29-operator-scripts\") pod \"keystone-db-create-h9dnk\" (UID: \"637ba10f-b8e0-48d9-9ea4-96569643bd29\") " pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.100810 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzp5t\" (UniqueName: \"kubernetes.io/projected/1d319fb8-91c8-4b37-9f5d-11df7656781e-kube-api-access-tzp5t\") pod \"keystone-1a4f-account-create-update-lcjf6\" (UID: \"1d319fb8-91c8-4b37-9f5d-11df7656781e\") " pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.100933 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d319fb8-91c8-4b37-9f5d-11df7656781e-operator-scripts\") pod \"keystone-1a4f-account-create-update-lcjf6\" (UID: \"1d319fb8-91c8-4b37-9f5d-11df7656781e\") " pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.100983 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdqbt\" (UniqueName: \"kubernetes.io/projected/637ba10f-b8e0-48d9-9ea4-96569643bd29-kube-api-access-vdqbt\") pod \"keystone-db-create-h9dnk\" (UID: \"637ba10f-b8e0-48d9-9ea4-96569643bd29\") " pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.102031 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/637ba10f-b8e0-48d9-9ea4-96569643bd29-operator-scripts\") pod \"keystone-db-create-h9dnk\" (UID: \"637ba10f-b8e0-48d9-9ea4-96569643bd29\") " pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.102208 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d319fb8-91c8-4b37-9f5d-11df7656781e-operator-scripts\") pod \"keystone-1a4f-account-create-update-lcjf6\" (UID: \"1d319fb8-91c8-4b37-9f5d-11df7656781e\") " pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.125287 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdqbt\" (UniqueName: \"kubernetes.io/projected/637ba10f-b8e0-48d9-9ea4-96569643bd29-kube-api-access-vdqbt\") pod \"keystone-db-create-h9dnk\" (UID: \"637ba10f-b8e0-48d9-9ea4-96569643bd29\") " pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.125466 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzp5t\" (UniqueName: \"kubernetes.io/projected/1d319fb8-91c8-4b37-9f5d-11df7656781e-kube-api-access-tzp5t\") pod \"keystone-1a4f-account-create-update-lcjf6\" (UID: \"1d319fb8-91c8-4b37-9f5d-11df7656781e\") " pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.198119 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.209311 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.352392 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="124214bc-17e4-4e8b-9c4b-1f52e0ccbd87" path="/var/lib/kubelet/pods/124214bc-17e4-4e8b-9c4b-1f52e0ccbd87/volumes" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.353961 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14645e66-70f7-411f-89cb-1cfabf743478" path="/var/lib/kubelet/pods/14645e66-70f7-411f-89cb-1cfabf743478/volumes" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.355180 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81fb977d-4fcb-4fde-9719-7d6d2b9dfae0" path="/var/lib/kubelet/pods/81fb977d-4fcb-4fde-9719-7d6d2b9dfae0/volumes" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.356158 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95197796-ac9e-47d1-b19f-92b534a42e5a" path="/var/lib/kubelet/pods/95197796-ac9e-47d1-b19f-92b534a42e5a/volumes" Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.691724 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6"] Dec 06 04:28:25 crc kubenswrapper[4718]: I1206 04:28:25.732901 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-h9dnk"] Dec 06 04:28:25 crc kubenswrapper[4718]: W1206 04:28:25.740115 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod637ba10f_b8e0_48d9_9ea4_96569643bd29.slice/crio-6dab16f753cc7add20b3f76e473c63bb97235ba9d04555f0d86434eb72be6018 WatchSource:0}: Error finding container 6dab16f753cc7add20b3f76e473c63bb97235ba9d04555f0d86434eb72be6018: Status 404 returned error can't find the container with id 6dab16f753cc7add20b3f76e473c63bb97235ba9d04555f0d86434eb72be6018 Dec 06 04:28:26 crc kubenswrapper[4718]: I1206 04:28:26.664803 4718 generic.go:334] "Generic (PLEG): container finished" podID="1d319fb8-91c8-4b37-9f5d-11df7656781e" containerID="ecdc7a98dc8744e7eee8630449d7484f2f721680c7499e37a24d19601ae47bd5" exitCode=0 Dec 06 04:28:26 crc kubenswrapper[4718]: I1206 04:28:26.664911 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" event={"ID":"1d319fb8-91c8-4b37-9f5d-11df7656781e","Type":"ContainerDied","Data":"ecdc7a98dc8744e7eee8630449d7484f2f721680c7499e37a24d19601ae47bd5"} Dec 06 04:28:26 crc kubenswrapper[4718]: I1206 04:28:26.665137 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" event={"ID":"1d319fb8-91c8-4b37-9f5d-11df7656781e","Type":"ContainerStarted","Data":"bda87c37d87020acc86cd587fca135e7c819d3b311436c0150598f28b095c762"} Dec 06 04:28:26 crc kubenswrapper[4718]: I1206 04:28:26.666974 4718 generic.go:334] "Generic (PLEG): container finished" podID="637ba10f-b8e0-48d9-9ea4-96569643bd29" containerID="bad8ba20b3c27557f13cd7a49404d358f44fc13fbe00db668711b3e7f2442a26" exitCode=0 Dec 06 04:28:26 crc kubenswrapper[4718]: I1206 04:28:26.667019 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-h9dnk" event={"ID":"637ba10f-b8e0-48d9-9ea4-96569643bd29","Type":"ContainerDied","Data":"bad8ba20b3c27557f13cd7a49404d358f44fc13fbe00db668711b3e7f2442a26"} Dec 06 04:28:26 crc kubenswrapper[4718]: I1206 04:28:26.667050 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-h9dnk" event={"ID":"637ba10f-b8e0-48d9-9ea4-96569643bd29","Type":"ContainerStarted","Data":"6dab16f753cc7add20b3f76e473c63bb97235ba9d04555f0d86434eb72be6018"} Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.043405 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.048647 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.148617 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d319fb8-91c8-4b37-9f5d-11df7656781e-operator-scripts\") pod \"1d319fb8-91c8-4b37-9f5d-11df7656781e\" (UID: \"1d319fb8-91c8-4b37-9f5d-11df7656781e\") " Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.148690 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzp5t\" (UniqueName: \"kubernetes.io/projected/1d319fb8-91c8-4b37-9f5d-11df7656781e-kube-api-access-tzp5t\") pod \"1d319fb8-91c8-4b37-9f5d-11df7656781e\" (UID: \"1d319fb8-91c8-4b37-9f5d-11df7656781e\") " Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.149743 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d319fb8-91c8-4b37-9f5d-11df7656781e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d319fb8-91c8-4b37-9f5d-11df7656781e" (UID: "1d319fb8-91c8-4b37-9f5d-11df7656781e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.153380 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d319fb8-91c8-4b37-9f5d-11df7656781e-kube-api-access-tzp5t" (OuterVolumeSpecName: "kube-api-access-tzp5t") pod "1d319fb8-91c8-4b37-9f5d-11df7656781e" (UID: "1d319fb8-91c8-4b37-9f5d-11df7656781e"). InnerVolumeSpecName "kube-api-access-tzp5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.249459 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdqbt\" (UniqueName: \"kubernetes.io/projected/637ba10f-b8e0-48d9-9ea4-96569643bd29-kube-api-access-vdqbt\") pod \"637ba10f-b8e0-48d9-9ea4-96569643bd29\" (UID: \"637ba10f-b8e0-48d9-9ea4-96569643bd29\") " Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.249557 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/637ba10f-b8e0-48d9-9ea4-96569643bd29-operator-scripts\") pod \"637ba10f-b8e0-48d9-9ea4-96569643bd29\" (UID: \"637ba10f-b8e0-48d9-9ea4-96569643bd29\") " Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.249821 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d319fb8-91c8-4b37-9f5d-11df7656781e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.249837 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzp5t\" (UniqueName: \"kubernetes.io/projected/1d319fb8-91c8-4b37-9f5d-11df7656781e-kube-api-access-tzp5t\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.250075 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637ba10f-b8e0-48d9-9ea4-96569643bd29-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "637ba10f-b8e0-48d9-9ea4-96569643bd29" (UID: "637ba10f-b8e0-48d9-9ea4-96569643bd29"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.252655 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/637ba10f-b8e0-48d9-9ea4-96569643bd29-kube-api-access-vdqbt" (OuterVolumeSpecName: "kube-api-access-vdqbt") pod "637ba10f-b8e0-48d9-9ea4-96569643bd29" (UID: "637ba10f-b8e0-48d9-9ea4-96569643bd29"). InnerVolumeSpecName "kube-api-access-vdqbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.351601 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdqbt\" (UniqueName: \"kubernetes.io/projected/637ba10f-b8e0-48d9-9ea4-96569643bd29-kube-api-access-vdqbt\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.352056 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/637ba10f-b8e0-48d9-9ea4-96569643bd29-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.685249 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" event={"ID":"1d319fb8-91c8-4b37-9f5d-11df7656781e","Type":"ContainerDied","Data":"bda87c37d87020acc86cd587fca135e7c819d3b311436c0150598f28b095c762"} Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.685283 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.685302 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bda87c37d87020acc86cd587fca135e7c819d3b311436c0150598f28b095c762" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.687454 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-h9dnk" event={"ID":"637ba10f-b8e0-48d9-9ea4-96569643bd29","Type":"ContainerDied","Data":"6dab16f753cc7add20b3f76e473c63bb97235ba9d04555f0d86434eb72be6018"} Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.687497 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dab16f753cc7add20b3f76e473c63bb97235ba9d04555f0d86434eb72be6018" Dec 06 04:28:28 crc kubenswrapper[4718]: I1206 04:28:28.687567 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-h9dnk" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.410399 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f94h5"] Dec 06 04:28:30 crc kubenswrapper[4718]: E1206 04:28:30.411190 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637ba10f-b8e0-48d9-9ea4-96569643bd29" containerName="mariadb-database-create" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.411208 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="637ba10f-b8e0-48d9-9ea4-96569643bd29" containerName="mariadb-database-create" Dec 06 04:28:30 crc kubenswrapper[4718]: E1206 04:28:30.411217 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d319fb8-91c8-4b37-9f5d-11df7656781e" containerName="mariadb-account-create-update" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.411226 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d319fb8-91c8-4b37-9f5d-11df7656781e" containerName="mariadb-account-create-update" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.411406 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d319fb8-91c8-4b37-9f5d-11df7656781e" containerName="mariadb-account-create-update" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.411415 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="637ba10f-b8e0-48d9-9ea4-96569643bd29" containerName="mariadb-database-create" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.411997 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.413680 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.414624 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.414924 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.415191 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-pjql7" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.426566 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f94h5"] Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.481870 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv7bx\" (UniqueName: \"kubernetes.io/projected/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-kube-api-access-lv7bx\") pod \"keystone-db-sync-f94h5\" (UID: \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\") " pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.482026 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-config-data\") pod \"keystone-db-sync-f94h5\" (UID: \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\") " pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.583553 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-config-data\") pod \"keystone-db-sync-f94h5\" (UID: \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\") " pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.583676 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv7bx\" (UniqueName: \"kubernetes.io/projected/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-kube-api-access-lv7bx\") pod \"keystone-db-sync-f94h5\" (UID: \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\") " pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.591457 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-config-data\") pod \"keystone-db-sync-f94h5\" (UID: \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\") " pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.618930 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv7bx\" (UniqueName: \"kubernetes.io/projected/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-kube-api-access-lv7bx\") pod \"keystone-db-sync-f94h5\" (UID: \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\") " pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:30 crc kubenswrapper[4718]: I1206 04:28:30.746749 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:31 crc kubenswrapper[4718]: I1206 04:28:31.036211 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f94h5"] Dec 06 04:28:31 crc kubenswrapper[4718]: W1206 04:28:31.044779 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fb8d08a_4afa_43f0_a0a2_891a5333d0dd.slice/crio-4b05bc8c187932aed977f1492ff0a3a680a49ccfba6cf93dc6339c45759bfa14 WatchSource:0}: Error finding container 4b05bc8c187932aed977f1492ff0a3a680a49ccfba6cf93dc6339c45759bfa14: Status 404 returned error can't find the container with id 4b05bc8c187932aed977f1492ff0a3a680a49ccfba6cf93dc6339c45759bfa14 Dec 06 04:28:31 crc kubenswrapper[4718]: I1206 04:28:31.717464 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-f94h5" event={"ID":"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd","Type":"ContainerStarted","Data":"a5c59c349ea3821949394f3f6698b50d0f399a9f755767d1d830b1b845c7e50f"} Dec 06 04:28:31 crc kubenswrapper[4718]: I1206 04:28:31.717514 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-f94h5" event={"ID":"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd","Type":"ContainerStarted","Data":"4b05bc8c187932aed977f1492ff0a3a680a49ccfba6cf93dc6339c45759bfa14"} Dec 06 04:28:32 crc kubenswrapper[4718]: I1206 04:28:32.726056 4718 generic.go:334] "Generic (PLEG): container finished" podID="2fb8d08a-4afa-43f0-a0a2-891a5333d0dd" containerID="a5c59c349ea3821949394f3f6698b50d0f399a9f755767d1d830b1b845c7e50f" exitCode=0 Dec 06 04:28:32 crc kubenswrapper[4718]: I1206 04:28:32.726143 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-f94h5" event={"ID":"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd","Type":"ContainerDied","Data":"a5c59c349ea3821949394f3f6698b50d0f399a9f755767d1d830b1b845c7e50f"} Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.010142 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.135129 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-config-data\") pod \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\" (UID: \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\") " Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.135272 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv7bx\" (UniqueName: \"kubernetes.io/projected/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-kube-api-access-lv7bx\") pod \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\" (UID: \"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd\") " Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.146276 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-kube-api-access-lv7bx" (OuterVolumeSpecName: "kube-api-access-lv7bx") pod "2fb8d08a-4afa-43f0-a0a2-891a5333d0dd" (UID: "2fb8d08a-4afa-43f0-a0a2-891a5333d0dd"). InnerVolumeSpecName "kube-api-access-lv7bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.195570 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-config-data" (OuterVolumeSpecName: "config-data") pod "2fb8d08a-4afa-43f0-a0a2-891a5333d0dd" (UID: "2fb8d08a-4afa-43f0-a0a2-891a5333d0dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.237105 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.237165 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv7bx\" (UniqueName: \"kubernetes.io/projected/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd-kube-api-access-lv7bx\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.744609 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-f94h5" event={"ID":"2fb8d08a-4afa-43f0-a0a2-891a5333d0dd","Type":"ContainerDied","Data":"4b05bc8c187932aed977f1492ff0a3a680a49ccfba6cf93dc6339c45759bfa14"} Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.744951 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b05bc8c187932aed977f1492ff0a3a680a49ccfba6cf93dc6339c45759bfa14" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.744656 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-f94h5" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.888053 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-5z5nq"] Dec 06 04:28:34 crc kubenswrapper[4718]: E1206 04:28:34.888346 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb8d08a-4afa-43f0-a0a2-891a5333d0dd" containerName="keystone-db-sync" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.888364 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb8d08a-4afa-43f0-a0a2-891a5333d0dd" containerName="keystone-db-sync" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.888519 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fb8d08a-4afa-43f0-a0a2-891a5333d0dd" containerName="keystone-db-sync" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.888972 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.892153 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.892367 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.892639 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-pjql7" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.892864 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.893926 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:28:34 crc kubenswrapper[4718]: I1206 04:28:34.903418 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-5z5nq"] Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.046716 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-fernet-keys\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.046784 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-credential-keys\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.046884 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25fww\" (UniqueName: \"kubernetes.io/projected/954d89dd-632f-49e8-9684-07f34440dc7b-kube-api-access-25fww\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.046945 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-config-data\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.046965 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-scripts\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.148177 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25fww\" (UniqueName: \"kubernetes.io/projected/954d89dd-632f-49e8-9684-07f34440dc7b-kube-api-access-25fww\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.148274 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-config-data\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.148307 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-scripts\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.148334 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-fernet-keys\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.148399 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-credential-keys\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.153507 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-config-data\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.157271 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-scripts\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.160133 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-fernet-keys\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.163771 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-credential-keys\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.176791 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25fww\" (UniqueName: \"kubernetes.io/projected/954d89dd-632f-49e8-9684-07f34440dc7b-kube-api-access-25fww\") pod \"keystone-bootstrap-5z5nq\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.207905 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.630803 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-5z5nq"] Dec 06 04:28:35 crc kubenswrapper[4718]: W1206 04:28:35.637410 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod954d89dd_632f_49e8_9684_07f34440dc7b.slice/crio-90c768cceefa74b993d70a32b56be69e4b117cd7f119ba2abc0aac0a9a344f93 WatchSource:0}: Error finding container 90c768cceefa74b993d70a32b56be69e4b117cd7f119ba2abc0aac0a9a344f93: Status 404 returned error can't find the container with id 90c768cceefa74b993d70a32b56be69e4b117cd7f119ba2abc0aac0a9a344f93 Dec 06 04:28:35 crc kubenswrapper[4718]: I1206 04:28:35.754293 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" event={"ID":"954d89dd-632f-49e8-9684-07f34440dc7b","Type":"ContainerStarted","Data":"90c768cceefa74b993d70a32b56be69e4b117cd7f119ba2abc0aac0a9a344f93"} Dec 06 04:28:36 crc kubenswrapper[4718]: I1206 04:28:36.763811 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" event={"ID":"954d89dd-632f-49e8-9684-07f34440dc7b","Type":"ContainerStarted","Data":"2675050ca4fc9fafac602f52eaf368505d0e1afcc10b9eebe2914b5d5ab1858f"} Dec 06 04:28:36 crc kubenswrapper[4718]: I1206 04:28:36.789221 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" podStartSLOduration=2.789009741 podStartE2EDuration="2.789009741s" podCreationTimestamp="2025-12-06 04:28:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:28:36.782978425 +0000 UTC m=+1305.788683596" watchObservedRunningTime="2025-12-06 04:28:36.789009741 +0000 UTC m=+1305.794714912" Dec 06 04:28:39 crc kubenswrapper[4718]: E1206 04:28:39.260834 4718 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod954d89dd_632f_49e8_9684_07f34440dc7b.slice/crio-conmon-2675050ca4fc9fafac602f52eaf368505d0e1afcc10b9eebe2914b5d5ab1858f.scope\": RecentStats: unable to find data in memory cache]" Dec 06 04:28:39 crc kubenswrapper[4718]: I1206 04:28:39.815022 4718 generic.go:334] "Generic (PLEG): container finished" podID="954d89dd-632f-49e8-9684-07f34440dc7b" containerID="2675050ca4fc9fafac602f52eaf368505d0e1afcc10b9eebe2914b5d5ab1858f" exitCode=0 Dec 06 04:28:39 crc kubenswrapper[4718]: I1206 04:28:39.815112 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" event={"ID":"954d89dd-632f-49e8-9684-07f34440dc7b","Type":"ContainerDied","Data":"2675050ca4fc9fafac602f52eaf368505d0e1afcc10b9eebe2914b5d5ab1858f"} Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.155347 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.243525 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-config-data\") pod \"954d89dd-632f-49e8-9684-07f34440dc7b\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.243599 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25fww\" (UniqueName: \"kubernetes.io/projected/954d89dd-632f-49e8-9684-07f34440dc7b-kube-api-access-25fww\") pod \"954d89dd-632f-49e8-9684-07f34440dc7b\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.243648 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-fernet-keys\") pod \"954d89dd-632f-49e8-9684-07f34440dc7b\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.243713 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-scripts\") pod \"954d89dd-632f-49e8-9684-07f34440dc7b\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.243734 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-credential-keys\") pod \"954d89dd-632f-49e8-9684-07f34440dc7b\" (UID: \"954d89dd-632f-49e8-9684-07f34440dc7b\") " Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.248951 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "954d89dd-632f-49e8-9684-07f34440dc7b" (UID: "954d89dd-632f-49e8-9684-07f34440dc7b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.248999 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "954d89dd-632f-49e8-9684-07f34440dc7b" (UID: "954d89dd-632f-49e8-9684-07f34440dc7b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.249191 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/954d89dd-632f-49e8-9684-07f34440dc7b-kube-api-access-25fww" (OuterVolumeSpecName: "kube-api-access-25fww") pod "954d89dd-632f-49e8-9684-07f34440dc7b" (UID: "954d89dd-632f-49e8-9684-07f34440dc7b"). InnerVolumeSpecName "kube-api-access-25fww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.249507 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-scripts" (OuterVolumeSpecName: "scripts") pod "954d89dd-632f-49e8-9684-07f34440dc7b" (UID: "954d89dd-632f-49e8-9684-07f34440dc7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.262511 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-config-data" (OuterVolumeSpecName: "config-data") pod "954d89dd-632f-49e8-9684-07f34440dc7b" (UID: "954d89dd-632f-49e8-9684-07f34440dc7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.346297 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.346349 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25fww\" (UniqueName: \"kubernetes.io/projected/954d89dd-632f-49e8-9684-07f34440dc7b-kube-api-access-25fww\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.346369 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.346384 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.346399 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/954d89dd-632f-49e8-9684-07f34440dc7b-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.836117 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" event={"ID":"954d89dd-632f-49e8-9684-07f34440dc7b","Type":"ContainerDied","Data":"90c768cceefa74b993d70a32b56be69e4b117cd7f119ba2abc0aac0a9a344f93"} Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.836174 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90c768cceefa74b993d70a32b56be69e4b117cd7f119ba2abc0aac0a9a344f93" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.836278 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-5z5nq" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.934948 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-796579d74d-p7pz5"] Dec 06 04:28:41 crc kubenswrapper[4718]: E1206 04:28:41.935380 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954d89dd-632f-49e8-9684-07f34440dc7b" containerName="keystone-bootstrap" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.935409 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="954d89dd-632f-49e8-9684-07f34440dc7b" containerName="keystone-bootstrap" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.935630 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="954d89dd-632f-49e8-9684-07f34440dc7b" containerName="keystone-bootstrap" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.936360 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.938421 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.938561 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.938771 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-pjql7" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.938871 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:28:41 crc kubenswrapper[4718]: I1206 04:28:41.956273 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-796579d74d-p7pz5"] Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.057520 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-config-data\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.057614 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-scripts\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.057726 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-credential-keys\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.057848 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j69m\" (UniqueName: \"kubernetes.io/projected/7b4acd03-36ae-49e5-863a-5442bbc869b2-kube-api-access-8j69m\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.057998 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-fernet-keys\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.159582 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-config-data\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.159650 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-scripts\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.159687 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-credential-keys\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.159730 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j69m\" (UniqueName: \"kubernetes.io/projected/7b4acd03-36ae-49e5-863a-5442bbc869b2-kube-api-access-8j69m\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.159755 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-fernet-keys\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.163392 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-config-data\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.163887 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-scripts\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.164192 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-fernet-keys\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.165321 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-credential-keys\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.191815 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j69m\" (UniqueName: \"kubernetes.io/projected/7b4acd03-36ae-49e5-863a-5442bbc869b2-kube-api-access-8j69m\") pod \"keystone-796579d74d-p7pz5\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.262425 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.482310 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-796579d74d-p7pz5"] Dec 06 04:28:42 crc kubenswrapper[4718]: I1206 04:28:42.849380 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" event={"ID":"7b4acd03-36ae-49e5-863a-5442bbc869b2","Type":"ContainerStarted","Data":"2a66b9c66f2434d78960addffa5b2168eb283ee1af0b8972746fccee63b941c9"} Dec 06 04:28:45 crc kubenswrapper[4718]: I1206 04:28:45.888417 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" event={"ID":"7b4acd03-36ae-49e5-863a-5442bbc869b2","Type":"ContainerStarted","Data":"e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4"} Dec 06 04:28:45 crc kubenswrapper[4718]: I1206 04:28:45.888827 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:28:45 crc kubenswrapper[4718]: I1206 04:28:45.906874 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" podStartSLOduration=4.90685523 podStartE2EDuration="4.90685523s" podCreationTimestamp="2025-12-06 04:28:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:28:45.903405625 +0000 UTC m=+1314.909110796" watchObservedRunningTime="2025-12-06 04:28:45.90685523 +0000 UTC m=+1314.912560391" Dec 06 04:29:13 crc kubenswrapper[4718]: I1206 04:29:13.712120 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.458392 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-5z5nq"] Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.473158 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-5z5nq"] Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.479830 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f94h5"] Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.485789 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f94h5"] Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.497443 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-796579d74d-p7pz5"] Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.497727 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" podUID="7b4acd03-36ae-49e5-863a-5442bbc869b2" containerName="keystone-api" containerID="cri-o://e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4" gracePeriod=30 Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.523896 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss"] Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.524908 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.531264 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss"] Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.687278 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbn46\" (UniqueName: \"kubernetes.io/projected/53ec40b0-88b9-46a4-a57a-1ba5245c0088-kube-api-access-rbn46\") pod \"keystone1a4f-account-delete-vk5ss\" (UID: \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\") " pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.687332 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec40b0-88b9-46a4-a57a-1ba5245c0088-operator-scripts\") pod \"keystone1a4f-account-delete-vk5ss\" (UID: \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\") " pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.788555 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbn46\" (UniqueName: \"kubernetes.io/projected/53ec40b0-88b9-46a4-a57a-1ba5245c0088-kube-api-access-rbn46\") pod \"keystone1a4f-account-delete-vk5ss\" (UID: \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\") " pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.788609 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec40b0-88b9-46a4-a57a-1ba5245c0088-operator-scripts\") pod \"keystone1a4f-account-delete-vk5ss\" (UID: \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\") " pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.789441 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec40b0-88b9-46a4-a57a-1ba5245c0088-operator-scripts\") pod \"keystone1a4f-account-delete-vk5ss\" (UID: \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\") " pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.811889 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbn46\" (UniqueName: \"kubernetes.io/projected/53ec40b0-88b9-46a4-a57a-1ba5245c0088-kube-api-access-rbn46\") pod \"keystone1a4f-account-delete-vk5ss\" (UID: \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\") " pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:29 crc kubenswrapper[4718]: I1206 04:29:29.843910 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:30 crc kubenswrapper[4718]: I1206 04:29:30.163600 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss"] Dec 06 04:29:30 crc kubenswrapper[4718]: I1206 04:29:30.265330 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" event={"ID":"53ec40b0-88b9-46a4-a57a-1ba5245c0088","Type":"ContainerStarted","Data":"12fc00092641304f260156239b41d8a0a9745ef4298de8499c66e44eb3db3453"} Dec 06 04:29:31 crc kubenswrapper[4718]: I1206 04:29:31.276332 4718 generic.go:334] "Generic (PLEG): container finished" podID="53ec40b0-88b9-46a4-a57a-1ba5245c0088" containerID="f89ac623a98973b93760a2f4df0bb78061ff34253647809e3821c0d65993aaba" exitCode=0 Dec 06 04:29:31 crc kubenswrapper[4718]: I1206 04:29:31.276386 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" event={"ID":"53ec40b0-88b9-46a4-a57a-1ba5245c0088","Type":"ContainerDied","Data":"f89ac623a98973b93760a2f4df0bb78061ff34253647809e3821c0d65993aaba"} Dec 06 04:29:31 crc kubenswrapper[4718]: I1206 04:29:31.341256 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fb8d08a-4afa-43f0-a0a2-891a5333d0dd" path="/var/lib/kubelet/pods/2fb8d08a-4afa-43f0-a0a2-891a5333d0dd/volumes" Dec 06 04:29:31 crc kubenswrapper[4718]: I1206 04:29:31.342229 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="954d89dd-632f-49e8-9684-07f34440dc7b" path="/var/lib/kubelet/pods/954d89dd-632f-49e8-9684-07f34440dc7b/volumes" Dec 06 04:29:32 crc kubenswrapper[4718]: I1206 04:29:32.657557 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:32 crc kubenswrapper[4718]: I1206 04:29:32.831781 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbn46\" (UniqueName: \"kubernetes.io/projected/53ec40b0-88b9-46a4-a57a-1ba5245c0088-kube-api-access-rbn46\") pod \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\" (UID: \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\") " Dec 06 04:29:32 crc kubenswrapper[4718]: I1206 04:29:32.831852 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec40b0-88b9-46a4-a57a-1ba5245c0088-operator-scripts\") pod \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\" (UID: \"53ec40b0-88b9-46a4-a57a-1ba5245c0088\") " Dec 06 04:29:32 crc kubenswrapper[4718]: I1206 04:29:32.832716 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53ec40b0-88b9-46a4-a57a-1ba5245c0088-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53ec40b0-88b9-46a4-a57a-1ba5245c0088" (UID: "53ec40b0-88b9-46a4-a57a-1ba5245c0088"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:29:32 crc kubenswrapper[4718]: I1206 04:29:32.849117 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53ec40b0-88b9-46a4-a57a-1ba5245c0088-kube-api-access-rbn46" (OuterVolumeSpecName: "kube-api-access-rbn46") pod "53ec40b0-88b9-46a4-a57a-1ba5245c0088" (UID: "53ec40b0-88b9-46a4-a57a-1ba5245c0088"). InnerVolumeSpecName "kube-api-access-rbn46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:29:32 crc kubenswrapper[4718]: I1206 04:29:32.933164 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbn46\" (UniqueName: \"kubernetes.io/projected/53ec40b0-88b9-46a4-a57a-1ba5245c0088-kube-api-access-rbn46\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:32 crc kubenswrapper[4718]: I1206 04:29:32.933225 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53ec40b0-88b9-46a4-a57a-1ba5245c0088-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:32 crc kubenswrapper[4718]: I1206 04:29:32.978964 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.135808 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j69m\" (UniqueName: \"kubernetes.io/projected/7b4acd03-36ae-49e5-863a-5442bbc869b2-kube-api-access-8j69m\") pod \"7b4acd03-36ae-49e5-863a-5442bbc869b2\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.135919 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-fernet-keys\") pod \"7b4acd03-36ae-49e5-863a-5442bbc869b2\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.135981 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-config-data\") pod \"7b4acd03-36ae-49e5-863a-5442bbc869b2\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.136036 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-scripts\") pod \"7b4acd03-36ae-49e5-863a-5442bbc869b2\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.136079 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-credential-keys\") pod \"7b4acd03-36ae-49e5-863a-5442bbc869b2\" (UID: \"7b4acd03-36ae-49e5-863a-5442bbc869b2\") " Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.138745 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b4acd03-36ae-49e5-863a-5442bbc869b2-kube-api-access-8j69m" (OuterVolumeSpecName: "kube-api-access-8j69m") pod "7b4acd03-36ae-49e5-863a-5442bbc869b2" (UID: "7b4acd03-36ae-49e5-863a-5442bbc869b2"). InnerVolumeSpecName "kube-api-access-8j69m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.139034 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7b4acd03-36ae-49e5-863a-5442bbc869b2" (UID: "7b4acd03-36ae-49e5-863a-5442bbc869b2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.140538 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7b4acd03-36ae-49e5-863a-5442bbc869b2" (UID: "7b4acd03-36ae-49e5-863a-5442bbc869b2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.140963 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-scripts" (OuterVolumeSpecName: "scripts") pod "7b4acd03-36ae-49e5-863a-5442bbc869b2" (UID: "7b4acd03-36ae-49e5-863a-5442bbc869b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.151271 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-config-data" (OuterVolumeSpecName: "config-data") pod "7b4acd03-36ae-49e5-863a-5442bbc869b2" (UID: "7b4acd03-36ae-49e5-863a-5442bbc869b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.237701 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.237737 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.237746 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.237755 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7b4acd03-36ae-49e5-863a-5442bbc869b2-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.237765 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j69m\" (UniqueName: \"kubernetes.io/projected/7b4acd03-36ae-49e5-863a-5442bbc869b2-kube-api-access-8j69m\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.300573 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" event={"ID":"53ec40b0-88b9-46a4-a57a-1ba5245c0088","Type":"ContainerDied","Data":"12fc00092641304f260156239b41d8a0a9745ef4298de8499c66e44eb3db3453"} Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.300616 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12fc00092641304f260156239b41d8a0a9745ef4298de8499c66e44eb3db3453" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.300629 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.302070 4718 generic.go:334] "Generic (PLEG): container finished" podID="7b4acd03-36ae-49e5-863a-5442bbc869b2" containerID="e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4" exitCode=0 Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.302104 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" event={"ID":"7b4acd03-36ae-49e5-863a-5442bbc869b2","Type":"ContainerDied","Data":"e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4"} Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.302124 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" event={"ID":"7b4acd03-36ae-49e5-863a-5442bbc869b2","Type":"ContainerDied","Data":"2a66b9c66f2434d78960addffa5b2168eb283ee1af0b8972746fccee63b941c9"} Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.302143 4718 scope.go:117] "RemoveContainer" containerID="e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.302282 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-796579d74d-p7pz5" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.338086 4718 scope.go:117] "RemoveContainer" containerID="e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4" Dec 06 04:29:33 crc kubenswrapper[4718]: E1206 04:29:33.338702 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4\": container with ID starting with e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4 not found: ID does not exist" containerID="e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.338754 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4"} err="failed to get container status \"e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4\": rpc error: code = NotFound desc = could not find container \"e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4\": container with ID starting with e93f81cc3382ee438f5d095aa79504154ff6191dcffe9f9ef7e36773377f08c4 not found: ID does not exist" Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.358658 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-796579d74d-p7pz5"] Dec 06 04:29:33 crc kubenswrapper[4718]: I1206 04:29:33.364570 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-796579d74d-p7pz5"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.562034 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-h9dnk"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.574470 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-h9dnk"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.592820 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.599844 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.605291 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone1a4f-account-delete-vk5ss"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.611574 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-1a4f-account-create-update-lcjf6"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.752877 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-k9mww"] Dec 06 04:29:34 crc kubenswrapper[4718]: E1206 04:29:34.753126 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ec40b0-88b9-46a4-a57a-1ba5245c0088" containerName="mariadb-account-delete" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.753164 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ec40b0-88b9-46a4-a57a-1ba5245c0088" containerName="mariadb-account-delete" Dec 06 04:29:34 crc kubenswrapper[4718]: E1206 04:29:34.753182 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4acd03-36ae-49e5-863a-5442bbc869b2" containerName="keystone-api" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.753188 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4acd03-36ae-49e5-863a-5442bbc869b2" containerName="keystone-api" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.753307 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b4acd03-36ae-49e5-863a-5442bbc869b2" containerName="keystone-api" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.753322 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="53ec40b0-88b9-46a4-a57a-1ba5245c0088" containerName="mariadb-account-delete" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.753710 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.766066 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-k9mww"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.774916 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.776222 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.779111 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.790261 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g"] Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.866921 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afa81b48-c76a-44ac-9770-94a17bc4e136-operator-scripts\") pod \"keystone-db-create-k9mww\" (UID: \"afa81b48-c76a-44ac-9770-94a17bc4e136\") " pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.867057 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-operator-scripts\") pod \"keystone-bbee-account-create-update-lx49g\" (UID: \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\") " pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.867108 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8pwv\" (UniqueName: \"kubernetes.io/projected/afa81b48-c76a-44ac-9770-94a17bc4e136-kube-api-access-q8pwv\") pod \"keystone-db-create-k9mww\" (UID: \"afa81b48-c76a-44ac-9770-94a17bc4e136\") " pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.867162 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvwx9\" (UniqueName: \"kubernetes.io/projected/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-kube-api-access-cvwx9\") pod \"keystone-bbee-account-create-update-lx49g\" (UID: \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\") " pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.968177 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-operator-scripts\") pod \"keystone-bbee-account-create-update-lx49g\" (UID: \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\") " pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.968343 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8pwv\" (UniqueName: \"kubernetes.io/projected/afa81b48-c76a-44ac-9770-94a17bc4e136-kube-api-access-q8pwv\") pod \"keystone-db-create-k9mww\" (UID: \"afa81b48-c76a-44ac-9770-94a17bc4e136\") " pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.968424 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvwx9\" (UniqueName: \"kubernetes.io/projected/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-kube-api-access-cvwx9\") pod \"keystone-bbee-account-create-update-lx49g\" (UID: \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\") " pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.968532 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afa81b48-c76a-44ac-9770-94a17bc4e136-operator-scripts\") pod \"keystone-db-create-k9mww\" (UID: \"afa81b48-c76a-44ac-9770-94a17bc4e136\") " pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.969268 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-operator-scripts\") pod \"keystone-bbee-account-create-update-lx49g\" (UID: \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\") " pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.969532 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afa81b48-c76a-44ac-9770-94a17bc4e136-operator-scripts\") pod \"keystone-db-create-k9mww\" (UID: \"afa81b48-c76a-44ac-9770-94a17bc4e136\") " pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.991313 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvwx9\" (UniqueName: \"kubernetes.io/projected/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-kube-api-access-cvwx9\") pod \"keystone-bbee-account-create-update-lx49g\" (UID: \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\") " pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:34 crc kubenswrapper[4718]: I1206 04:29:34.999805 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8pwv\" (UniqueName: \"kubernetes.io/projected/afa81b48-c76a-44ac-9770-94a17bc4e136-kube-api-access-q8pwv\") pod \"keystone-db-create-k9mww\" (UID: \"afa81b48-c76a-44ac-9770-94a17bc4e136\") " pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:35 crc kubenswrapper[4718]: I1206 04:29:35.069885 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:35 crc kubenswrapper[4718]: I1206 04:29:35.095469 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:35 crc kubenswrapper[4718]: I1206 04:29:35.338113 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d319fb8-91c8-4b37-9f5d-11df7656781e" path="/var/lib/kubelet/pods/1d319fb8-91c8-4b37-9f5d-11df7656781e/volumes" Dec 06 04:29:35 crc kubenswrapper[4718]: I1206 04:29:35.338992 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53ec40b0-88b9-46a4-a57a-1ba5245c0088" path="/var/lib/kubelet/pods/53ec40b0-88b9-46a4-a57a-1ba5245c0088/volumes" Dec 06 04:29:35 crc kubenswrapper[4718]: I1206 04:29:35.339712 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="637ba10f-b8e0-48d9-9ea4-96569643bd29" path="/var/lib/kubelet/pods/637ba10f-b8e0-48d9-9ea4-96569643bd29/volumes" Dec 06 04:29:35 crc kubenswrapper[4718]: I1206 04:29:35.340380 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b4acd03-36ae-49e5-863a-5442bbc869b2" path="/var/lib/kubelet/pods/7b4acd03-36ae-49e5-863a-5442bbc869b2/volumes" Dec 06 04:29:35 crc kubenswrapper[4718]: I1206 04:29:35.532412 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-k9mww"] Dec 06 04:29:35 crc kubenswrapper[4718]: I1206 04:29:35.557211 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g"] Dec 06 04:29:35 crc kubenswrapper[4718]: W1206 04:29:35.561837 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf13aaefa_a7d6_4ac6_b67f_1710f79b1791.slice/crio-012eb2171528b34a4b57638c57ff05137b669a166458edf818ba33c43b0fe8de WatchSource:0}: Error finding container 012eb2171528b34a4b57638c57ff05137b669a166458edf818ba33c43b0fe8de: Status 404 returned error can't find the container with id 012eb2171528b34a4b57638c57ff05137b669a166458edf818ba33c43b0fe8de Dec 06 04:29:36 crc kubenswrapper[4718]: I1206 04:29:36.324952 4718 generic.go:334] "Generic (PLEG): container finished" podID="f13aaefa-a7d6-4ac6-b67f-1710f79b1791" containerID="5f29c1709190ab16ceb77a4fdb2742e157f5612881c1d659cc9bb610df21193a" exitCode=0 Dec 06 04:29:36 crc kubenswrapper[4718]: I1206 04:29:36.325027 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" event={"ID":"f13aaefa-a7d6-4ac6-b67f-1710f79b1791","Type":"ContainerDied","Data":"5f29c1709190ab16ceb77a4fdb2742e157f5612881c1d659cc9bb610df21193a"} Dec 06 04:29:36 crc kubenswrapper[4718]: I1206 04:29:36.325418 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" event={"ID":"f13aaefa-a7d6-4ac6-b67f-1710f79b1791","Type":"ContainerStarted","Data":"012eb2171528b34a4b57638c57ff05137b669a166458edf818ba33c43b0fe8de"} Dec 06 04:29:36 crc kubenswrapper[4718]: I1206 04:29:36.326807 4718 generic.go:334] "Generic (PLEG): container finished" podID="afa81b48-c76a-44ac-9770-94a17bc4e136" containerID="1abd178aced5617cc967f88b78ccca1c13cb4b851d0855d6fe169e8c22de8027" exitCode=0 Dec 06 04:29:36 crc kubenswrapper[4718]: I1206 04:29:36.326834 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-k9mww" event={"ID":"afa81b48-c76a-44ac-9770-94a17bc4e136","Type":"ContainerDied","Data":"1abd178aced5617cc967f88b78ccca1c13cb4b851d0855d6fe169e8c22de8027"} Dec 06 04:29:36 crc kubenswrapper[4718]: I1206 04:29:36.326853 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-k9mww" event={"ID":"afa81b48-c76a-44ac-9770-94a17bc4e136","Type":"ContainerStarted","Data":"224e27ac1326d184c71abcbe3a26e5fe45a76a19cbe1c531f8f0007e38567656"} Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.696080 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.702808 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.724783 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8pwv\" (UniqueName: \"kubernetes.io/projected/afa81b48-c76a-44ac-9770-94a17bc4e136-kube-api-access-q8pwv\") pod \"afa81b48-c76a-44ac-9770-94a17bc4e136\" (UID: \"afa81b48-c76a-44ac-9770-94a17bc4e136\") " Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.726257 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-operator-scripts\") pod \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\" (UID: \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\") " Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.726340 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afa81b48-c76a-44ac-9770-94a17bc4e136-operator-scripts\") pod \"afa81b48-c76a-44ac-9770-94a17bc4e136\" (UID: \"afa81b48-c76a-44ac-9770-94a17bc4e136\") " Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.726374 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvwx9\" (UniqueName: \"kubernetes.io/projected/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-kube-api-access-cvwx9\") pod \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\" (UID: \"f13aaefa-a7d6-4ac6-b67f-1710f79b1791\") " Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.726904 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f13aaefa-a7d6-4ac6-b67f-1710f79b1791" (UID: "f13aaefa-a7d6-4ac6-b67f-1710f79b1791"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.727258 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa81b48-c76a-44ac-9770-94a17bc4e136-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "afa81b48-c76a-44ac-9770-94a17bc4e136" (UID: "afa81b48-c76a-44ac-9770-94a17bc4e136"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.732138 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afa81b48-c76a-44ac-9770-94a17bc4e136-kube-api-access-q8pwv" (OuterVolumeSpecName: "kube-api-access-q8pwv") pod "afa81b48-c76a-44ac-9770-94a17bc4e136" (UID: "afa81b48-c76a-44ac-9770-94a17bc4e136"). InnerVolumeSpecName "kube-api-access-q8pwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.740358 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-kube-api-access-cvwx9" (OuterVolumeSpecName: "kube-api-access-cvwx9") pod "f13aaefa-a7d6-4ac6-b67f-1710f79b1791" (UID: "f13aaefa-a7d6-4ac6-b67f-1710f79b1791"). InnerVolumeSpecName "kube-api-access-cvwx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.827302 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.827340 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afa81b48-c76a-44ac-9770-94a17bc4e136-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.827373 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvwx9\" (UniqueName: \"kubernetes.io/projected/f13aaefa-a7d6-4ac6-b67f-1710f79b1791-kube-api-access-cvwx9\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:37 crc kubenswrapper[4718]: I1206 04:29:37.827386 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8pwv\" (UniqueName: \"kubernetes.io/projected/afa81b48-c76a-44ac-9770-94a17bc4e136-kube-api-access-q8pwv\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:38 crc kubenswrapper[4718]: I1206 04:29:38.344210 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-k9mww" event={"ID":"afa81b48-c76a-44ac-9770-94a17bc4e136","Type":"ContainerDied","Data":"224e27ac1326d184c71abcbe3a26e5fe45a76a19cbe1c531f8f0007e38567656"} Dec 06 04:29:38 crc kubenswrapper[4718]: I1206 04:29:38.344257 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-k9mww" Dec 06 04:29:38 crc kubenswrapper[4718]: I1206 04:29:38.344275 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="224e27ac1326d184c71abcbe3a26e5fe45a76a19cbe1c531f8f0007e38567656" Dec 06 04:29:38 crc kubenswrapper[4718]: I1206 04:29:38.345716 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" event={"ID":"f13aaefa-a7d6-4ac6-b67f-1710f79b1791","Type":"ContainerDied","Data":"012eb2171528b34a4b57638c57ff05137b669a166458edf818ba33c43b0fe8de"} Dec 06 04:29:38 crc kubenswrapper[4718]: I1206 04:29:38.345741 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="012eb2171528b34a4b57638c57ff05137b669a166458edf818ba33c43b0fe8de" Dec 06 04:29:38 crc kubenswrapper[4718]: I1206 04:29:38.345784 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.326525 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-zzlxc"] Dec 06 04:29:40 crc kubenswrapper[4718]: E1206 04:29:40.327053 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f13aaefa-a7d6-4ac6-b67f-1710f79b1791" containerName="mariadb-account-create-update" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.327066 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f13aaefa-a7d6-4ac6-b67f-1710f79b1791" containerName="mariadb-account-create-update" Dec 06 04:29:40 crc kubenswrapper[4718]: E1206 04:29:40.327079 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa81b48-c76a-44ac-9770-94a17bc4e136" containerName="mariadb-database-create" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.327086 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa81b48-c76a-44ac-9770-94a17bc4e136" containerName="mariadb-database-create" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.327203 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f13aaefa-a7d6-4ac6-b67f-1710f79b1791" containerName="mariadb-account-create-update" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.327214 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="afa81b48-c76a-44ac-9770-94a17bc4e136" containerName="mariadb-database-create" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.327651 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.330469 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-7zvlm" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.330702 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.330703 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.332089 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.349999 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-zzlxc"] Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.469593 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8thfg\" (UniqueName: \"kubernetes.io/projected/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-kube-api-access-8thfg\") pod \"keystone-db-sync-zzlxc\" (UID: \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\") " pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.469715 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-config-data\") pod \"keystone-db-sync-zzlxc\" (UID: \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\") " pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.571666 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-config-data\") pod \"keystone-db-sync-zzlxc\" (UID: \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\") " pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.571874 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8thfg\" (UniqueName: \"kubernetes.io/projected/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-kube-api-access-8thfg\") pod \"keystone-db-sync-zzlxc\" (UID: \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\") " pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.589212 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-config-data\") pod \"keystone-db-sync-zzlxc\" (UID: \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\") " pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.602828 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8thfg\" (UniqueName: \"kubernetes.io/projected/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-kube-api-access-8thfg\") pod \"keystone-db-sync-zzlxc\" (UID: \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\") " pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:40 crc kubenswrapper[4718]: I1206 04:29:40.646546 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:41 crc kubenswrapper[4718]: I1206 04:29:41.087210 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-zzlxc"] Dec 06 04:29:41 crc kubenswrapper[4718]: I1206 04:29:41.385205 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" event={"ID":"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2","Type":"ContainerStarted","Data":"092cd57e8bc8cf4daeb067fcee66056748f8abaeb12ac27e013d5ab053e98444"} Dec 06 04:29:42 crc kubenswrapper[4718]: I1206 04:29:42.402080 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" event={"ID":"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2","Type":"ContainerStarted","Data":"440203bc8764d63be0160e90ad48f81dee27949f692af3dc35136f88363a20e9"} Dec 06 04:29:42 crc kubenswrapper[4718]: I1206 04:29:42.423470 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" podStartSLOduration=2.423451299 podStartE2EDuration="2.423451299s" podCreationTimestamp="2025-12-06 04:29:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:29:42.421750593 +0000 UTC m=+1371.427455804" watchObservedRunningTime="2025-12-06 04:29:42.423451299 +0000 UTC m=+1371.429156470" Dec 06 04:29:43 crc kubenswrapper[4718]: I1206 04:29:43.412332 4718 generic.go:334] "Generic (PLEG): container finished" podID="5cf00cd1-3b58-4426-9d60-b5e16c2cafa2" containerID="440203bc8764d63be0160e90ad48f81dee27949f692af3dc35136f88363a20e9" exitCode=0 Dec 06 04:29:43 crc kubenswrapper[4718]: I1206 04:29:43.412382 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" event={"ID":"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2","Type":"ContainerDied","Data":"440203bc8764d63be0160e90ad48f81dee27949f692af3dc35136f88363a20e9"} Dec 06 04:29:44 crc kubenswrapper[4718]: I1206 04:29:44.741110 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:44 crc kubenswrapper[4718]: I1206 04:29:44.945519 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8thfg\" (UniqueName: \"kubernetes.io/projected/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-kube-api-access-8thfg\") pod \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\" (UID: \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\") " Dec 06 04:29:44 crc kubenswrapper[4718]: I1206 04:29:44.945772 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-config-data\") pod \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\" (UID: \"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2\") " Dec 06 04:29:44 crc kubenswrapper[4718]: I1206 04:29:44.962153 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-kube-api-access-8thfg" (OuterVolumeSpecName: "kube-api-access-8thfg") pod "5cf00cd1-3b58-4426-9d60-b5e16c2cafa2" (UID: "5cf00cd1-3b58-4426-9d60-b5e16c2cafa2"). InnerVolumeSpecName "kube-api-access-8thfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.002503 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-config-data" (OuterVolumeSpecName: "config-data") pod "5cf00cd1-3b58-4426-9d60-b5e16c2cafa2" (UID: "5cf00cd1-3b58-4426-9d60-b5e16c2cafa2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.047387 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.047426 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8thfg\" (UniqueName: \"kubernetes.io/projected/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2-kube-api-access-8thfg\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.434288 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" event={"ID":"5cf00cd1-3b58-4426-9d60-b5e16c2cafa2","Type":"ContainerDied","Data":"092cd57e8bc8cf4daeb067fcee66056748f8abaeb12ac27e013d5ab053e98444"} Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.434356 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-zzlxc" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.434368 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="092cd57e8bc8cf4daeb067fcee66056748f8abaeb12ac27e013d5ab053e98444" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.635285 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-g7vxp"] Dec 06 04:29:45 crc kubenswrapper[4718]: E1206 04:29:45.635655 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf00cd1-3b58-4426-9d60-b5e16c2cafa2" containerName="keystone-db-sync" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.635679 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf00cd1-3b58-4426-9d60-b5e16c2cafa2" containerName="keystone-db-sync" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.635871 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cf00cd1-3b58-4426-9d60-b5e16c2cafa2" containerName="keystone-db-sync" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.636643 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.643055 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.643591 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.643732 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-7zvlm" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.643938 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.643939 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.665449 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-g7vxp"] Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.760380 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lvtc\" (UniqueName: \"kubernetes.io/projected/6e15091c-5c28-411a-906d-31e3860fcfe4-kube-api-access-4lvtc\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.760435 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-credential-keys\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.760464 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-config-data\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.760494 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-scripts\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.760533 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-fernet-keys\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.861713 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-scripts\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.861781 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-fernet-keys\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.861884 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lvtc\" (UniqueName: \"kubernetes.io/projected/6e15091c-5c28-411a-906d-31e3860fcfe4-kube-api-access-4lvtc\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.861909 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-credential-keys\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.861939 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-config-data\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.866465 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-fernet-keys\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.867075 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-scripts\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.868206 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-credential-keys\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.869996 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-config-data\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.882858 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lvtc\" (UniqueName: \"kubernetes.io/projected/6e15091c-5c28-411a-906d-31e3860fcfe4-kube-api-access-4lvtc\") pod \"keystone-bootstrap-g7vxp\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:45 crc kubenswrapper[4718]: I1206 04:29:45.967066 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:46 crc kubenswrapper[4718]: I1206 04:29:46.410712 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-g7vxp"] Dec 06 04:29:46 crc kubenswrapper[4718]: I1206 04:29:46.446911 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" event={"ID":"6e15091c-5c28-411a-906d-31e3860fcfe4","Type":"ContainerStarted","Data":"0d8ec62ff52f9029f816f412716756995ccc553f6026a1c1c3a4ae5eb152b719"} Dec 06 04:29:47 crc kubenswrapper[4718]: I1206 04:29:47.456623 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" event={"ID":"6e15091c-5c28-411a-906d-31e3860fcfe4","Type":"ContainerStarted","Data":"edb179856b3c24498b6199c969cc70cd97420e32028272d18b4c4779b076904f"} Dec 06 04:29:47 crc kubenswrapper[4718]: I1206 04:29:47.477063 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" podStartSLOduration=2.477044903 podStartE2EDuration="2.477044903s" podCreationTimestamp="2025-12-06 04:29:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:29:47.474476244 +0000 UTC m=+1376.480181445" watchObservedRunningTime="2025-12-06 04:29:47.477044903 +0000 UTC m=+1376.482750074" Dec 06 04:29:49 crc kubenswrapper[4718]: I1206 04:29:49.479339 4718 generic.go:334] "Generic (PLEG): container finished" podID="6e15091c-5c28-411a-906d-31e3860fcfe4" containerID="edb179856b3c24498b6199c969cc70cd97420e32028272d18b4c4779b076904f" exitCode=0 Dec 06 04:29:49 crc kubenswrapper[4718]: I1206 04:29:49.479450 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" event={"ID":"6e15091c-5c28-411a-906d-31e3860fcfe4","Type":"ContainerDied","Data":"edb179856b3c24498b6199c969cc70cd97420e32028272d18b4c4779b076904f"} Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.783076 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.944175 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lvtc\" (UniqueName: \"kubernetes.io/projected/6e15091c-5c28-411a-906d-31e3860fcfe4-kube-api-access-4lvtc\") pod \"6e15091c-5c28-411a-906d-31e3860fcfe4\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.944255 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-config-data\") pod \"6e15091c-5c28-411a-906d-31e3860fcfe4\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.944306 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-credential-keys\") pod \"6e15091c-5c28-411a-906d-31e3860fcfe4\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.944378 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-scripts\") pod \"6e15091c-5c28-411a-906d-31e3860fcfe4\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.944443 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-fernet-keys\") pod \"6e15091c-5c28-411a-906d-31e3860fcfe4\" (UID: \"6e15091c-5c28-411a-906d-31e3860fcfe4\") " Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.956144 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6e15091c-5c28-411a-906d-31e3860fcfe4" (UID: "6e15091c-5c28-411a-906d-31e3860fcfe4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.970383 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-scripts" (OuterVolumeSpecName: "scripts") pod "6e15091c-5c28-411a-906d-31e3860fcfe4" (UID: "6e15091c-5c28-411a-906d-31e3860fcfe4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.970465 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6e15091c-5c28-411a-906d-31e3860fcfe4" (UID: "6e15091c-5c28-411a-906d-31e3860fcfe4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:29:50 crc kubenswrapper[4718]: I1206 04:29:50.972393 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e15091c-5c28-411a-906d-31e3860fcfe4-kube-api-access-4lvtc" (OuterVolumeSpecName: "kube-api-access-4lvtc") pod "6e15091c-5c28-411a-906d-31e3860fcfe4" (UID: "6e15091c-5c28-411a-906d-31e3860fcfe4"). InnerVolumeSpecName "kube-api-access-4lvtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.029463 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-config-data" (OuterVolumeSpecName: "config-data") pod "6e15091c-5c28-411a-906d-31e3860fcfe4" (UID: "6e15091c-5c28-411a-906d-31e3860fcfe4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.046301 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lvtc\" (UniqueName: \"kubernetes.io/projected/6e15091c-5c28-411a-906d-31e3860fcfe4-kube-api-access-4lvtc\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.046343 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.046357 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.046369 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.046381 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e15091c-5c28-411a-906d-31e3860fcfe4-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.497351 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" event={"ID":"6e15091c-5c28-411a-906d-31e3860fcfe4","Type":"ContainerDied","Data":"0d8ec62ff52f9029f816f412716756995ccc553f6026a1c1c3a4ae5eb152b719"} Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.497741 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d8ec62ff52f9029f816f412716756995ccc553f6026a1c1c3a4ae5eb152b719" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.497431 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-g7vxp" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.596802 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2"] Dec 06 04:29:51 crc kubenswrapper[4718]: E1206 04:29:51.597223 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e15091c-5c28-411a-906d-31e3860fcfe4" containerName="keystone-bootstrap" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.597275 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e15091c-5c28-411a-906d-31e3860fcfe4" containerName="keystone-bootstrap" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.597467 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e15091c-5c28-411a-906d-31e3860fcfe4" containerName="keystone-bootstrap" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.598062 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.601144 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.601695 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-7zvlm" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.602001 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.602299 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.619265 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2"] Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.758044 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-fernet-keys\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.758286 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-scripts\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.758369 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-config-data\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.758417 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52b8d\" (UniqueName: \"kubernetes.io/projected/5549ac76-32a3-4ec8-a924-df653c1bf781-kube-api-access-52b8d\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.758469 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-credential-keys\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.859475 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-scripts\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.859600 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-config-data\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.859646 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52b8d\" (UniqueName: \"kubernetes.io/projected/5549ac76-32a3-4ec8-a924-df653c1bf781-kube-api-access-52b8d\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.859696 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-credential-keys\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.859809 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-fernet-keys\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.866804 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-fernet-keys\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.869496 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-config-data\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.871154 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-scripts\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.878562 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-credential-keys\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.894844 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52b8d\" (UniqueName: \"kubernetes.io/projected/5549ac76-32a3-4ec8-a924-df653c1bf781-kube-api-access-52b8d\") pod \"keystone-855fbcf9b4-f8gv2\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:51 crc kubenswrapper[4718]: I1206 04:29:51.914129 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:53 crc kubenswrapper[4718]: I1206 04:29:53.169715 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2"] Dec 06 04:29:53 crc kubenswrapper[4718]: I1206 04:29:53.511673 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" event={"ID":"5549ac76-32a3-4ec8-a924-df653c1bf781","Type":"ContainerStarted","Data":"8e632837865cd0d9e76c72b29ce8781b4510667e413fac0d212f2b8cdddf3f1e"} Dec 06 04:29:53 crc kubenswrapper[4718]: I1206 04:29:53.512088 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:29:53 crc kubenswrapper[4718]: I1206 04:29:53.512113 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" event={"ID":"5549ac76-32a3-4ec8-a924-df653c1bf781","Type":"ContainerStarted","Data":"18b0e1b59693d54b8a6eadd4dd8473af52b2d54111de149ced1216e82c0fc535"} Dec 06 04:29:53 crc kubenswrapper[4718]: I1206 04:29:53.537307 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" podStartSLOduration=2.537293045 podStartE2EDuration="2.537293045s" podCreationTimestamp="2025-12-06 04:29:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:29:53.532357984 +0000 UTC m=+1382.538063155" watchObservedRunningTime="2025-12-06 04:29:53.537293045 +0000 UTC m=+1382.542998206" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.141573 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8"] Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.143221 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.146414 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.146810 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.156952 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8"] Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.287218 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h7gp\" (UniqueName: \"kubernetes.io/projected/95bd0038-6029-4bd5-aa55-34d50b54010b-kube-api-access-2h7gp\") pod \"collect-profiles-29416590-9h7r8\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.287371 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95bd0038-6029-4bd5-aa55-34d50b54010b-config-volume\") pod \"collect-profiles-29416590-9h7r8\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.287399 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95bd0038-6029-4bd5-aa55-34d50b54010b-secret-volume\") pod \"collect-profiles-29416590-9h7r8\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.388954 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95bd0038-6029-4bd5-aa55-34d50b54010b-secret-volume\") pod \"collect-profiles-29416590-9h7r8\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.389154 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h7gp\" (UniqueName: \"kubernetes.io/projected/95bd0038-6029-4bd5-aa55-34d50b54010b-kube-api-access-2h7gp\") pod \"collect-profiles-29416590-9h7r8\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.389349 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95bd0038-6029-4bd5-aa55-34d50b54010b-config-volume\") pod \"collect-profiles-29416590-9h7r8\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.391067 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95bd0038-6029-4bd5-aa55-34d50b54010b-config-volume\") pod \"collect-profiles-29416590-9h7r8\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.401435 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95bd0038-6029-4bd5-aa55-34d50b54010b-secret-volume\") pod \"collect-profiles-29416590-9h7r8\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.446377 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h7gp\" (UniqueName: \"kubernetes.io/projected/95bd0038-6029-4bd5-aa55-34d50b54010b-kube-api-access-2h7gp\") pod \"collect-profiles-29416590-9h7r8\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.478945 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:00 crc kubenswrapper[4718]: I1206 04:30:00.891704 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8"] Dec 06 04:30:00 crc kubenswrapper[4718]: W1206 04:30:00.895530 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95bd0038_6029_4bd5_aa55_34d50b54010b.slice/crio-200a9cf15c430ed060c90edd58db693254da2ae552c2a2e2e157300bde6d97d4 WatchSource:0}: Error finding container 200a9cf15c430ed060c90edd58db693254da2ae552c2a2e2e157300bde6d97d4: Status 404 returned error can't find the container with id 200a9cf15c430ed060c90edd58db693254da2ae552c2a2e2e157300bde6d97d4 Dec 06 04:30:01 crc kubenswrapper[4718]: I1206 04:30:01.584047 4718 generic.go:334] "Generic (PLEG): container finished" podID="95bd0038-6029-4bd5-aa55-34d50b54010b" containerID="d195f1bf7e53599e689c30a7ab74ffd15785a543d1457539b9e8c567e5ccc1a8" exitCode=0 Dec 06 04:30:01 crc kubenswrapper[4718]: I1206 04:30:01.584121 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" event={"ID":"95bd0038-6029-4bd5-aa55-34d50b54010b","Type":"ContainerDied","Data":"d195f1bf7e53599e689c30a7ab74ffd15785a543d1457539b9e8c567e5ccc1a8"} Dec 06 04:30:01 crc kubenswrapper[4718]: I1206 04:30:01.584187 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" event={"ID":"95bd0038-6029-4bd5-aa55-34d50b54010b","Type":"ContainerStarted","Data":"200a9cf15c430ed060c90edd58db693254da2ae552c2a2e2e157300bde6d97d4"} Dec 06 04:30:02 crc kubenswrapper[4718]: I1206 04:30:02.860459 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.027952 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95bd0038-6029-4bd5-aa55-34d50b54010b-secret-volume\") pod \"95bd0038-6029-4bd5-aa55-34d50b54010b\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.028144 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95bd0038-6029-4bd5-aa55-34d50b54010b-config-volume\") pod \"95bd0038-6029-4bd5-aa55-34d50b54010b\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.028214 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h7gp\" (UniqueName: \"kubernetes.io/projected/95bd0038-6029-4bd5-aa55-34d50b54010b-kube-api-access-2h7gp\") pod \"95bd0038-6029-4bd5-aa55-34d50b54010b\" (UID: \"95bd0038-6029-4bd5-aa55-34d50b54010b\") " Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.028611 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95bd0038-6029-4bd5-aa55-34d50b54010b-config-volume" (OuterVolumeSpecName: "config-volume") pod "95bd0038-6029-4bd5-aa55-34d50b54010b" (UID: "95bd0038-6029-4bd5-aa55-34d50b54010b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.028779 4718 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95bd0038-6029-4bd5-aa55-34d50b54010b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.034008 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95bd0038-6029-4bd5-aa55-34d50b54010b-kube-api-access-2h7gp" (OuterVolumeSpecName: "kube-api-access-2h7gp") pod "95bd0038-6029-4bd5-aa55-34d50b54010b" (UID: "95bd0038-6029-4bd5-aa55-34d50b54010b"). InnerVolumeSpecName "kube-api-access-2h7gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.034512 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95bd0038-6029-4bd5-aa55-34d50b54010b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "95bd0038-6029-4bd5-aa55-34d50b54010b" (UID: "95bd0038-6029-4bd5-aa55-34d50b54010b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.131295 4718 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95bd0038-6029-4bd5-aa55-34d50b54010b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.131345 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h7gp\" (UniqueName: \"kubernetes.io/projected/95bd0038-6029-4bd5-aa55-34d50b54010b-kube-api-access-2h7gp\") on node \"crc\" DevicePath \"\"" Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.601138 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" event={"ID":"95bd0038-6029-4bd5-aa55-34d50b54010b","Type":"ContainerDied","Data":"200a9cf15c430ed060c90edd58db693254da2ae552c2a2e2e157300bde6d97d4"} Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.601854 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="200a9cf15c430ed060c90edd58db693254da2ae552c2a2e2e157300bde6d97d4" Dec 06 04:30:03 crc kubenswrapper[4718]: I1206 04:30:03.601277 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416590-9h7r8" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.317655 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.635849 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/openstackclient"] Dec 06 04:30:23 crc kubenswrapper[4718]: E1206 04:30:23.636075 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95bd0038-6029-4bd5-aa55-34d50b54010b" containerName="collect-profiles" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.636119 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="95bd0038-6029-4bd5-aa55-34d50b54010b" containerName="collect-profiles" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.636271 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="95bd0038-6029-4bd5-aa55-34d50b54010b" containerName="collect-profiles" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.636663 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.639041 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"openstack-config" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.639859 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"default-dockercfg-5v6xt" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.641359 4718 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"openstack-config-secret" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.647539 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.792885 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config\") pod \"openstackclient\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.792939 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config-secret\") pod \"openstackclient\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.793157 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpgs9\" (UniqueName: \"kubernetes.io/projected/a7582aad-5aa4-494b-84c4-a164850e254f-kube-api-access-xpgs9\") pod \"openstackclient\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.894872 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config\") pod \"openstackclient\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.894949 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config-secret\") pod \"openstackclient\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.895051 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpgs9\" (UniqueName: \"kubernetes.io/projected/a7582aad-5aa4-494b-84c4-a164850e254f-kube-api-access-xpgs9\") pod \"openstackclient\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.896401 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config\") pod \"openstackclient\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.904393 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config-secret\") pod \"openstackclient\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.913699 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpgs9\" (UniqueName: \"kubernetes.io/projected/a7582aad-5aa4-494b-84c4-a164850e254f-kube-api-access-xpgs9\") pod \"openstackclient\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:23 crc kubenswrapper[4718]: I1206 04:30:23.961036 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstackclient" Dec 06 04:30:24 crc kubenswrapper[4718]: I1206 04:30:24.213936 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Dec 06 04:30:24 crc kubenswrapper[4718]: I1206 04:30:24.221848 4718 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:30:24 crc kubenswrapper[4718]: I1206 04:30:24.769985 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstackclient" event={"ID":"a7582aad-5aa4-494b-84c4-a164850e254f","Type":"ContainerStarted","Data":"6cf83db62fc8be6d4aaaaaf62e4f6e19468d4ca61247dd64a955829282cbad92"} Dec 06 04:30:27 crc kubenswrapper[4718]: I1206 04:30:27.877190 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:30:27 crc kubenswrapper[4718]: I1206 04:30:27.877303 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:30:33 crc kubenswrapper[4718]: I1206 04:30:33.865088 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstackclient" event={"ID":"a7582aad-5aa4-494b-84c4-a164850e254f","Type":"ContainerStarted","Data":"38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14"} Dec 06 04:30:33 crc kubenswrapper[4718]: I1206 04:30:33.894196 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/openstackclient" podStartSLOduration=2.215579215 podStartE2EDuration="10.894167437s" podCreationTimestamp="2025-12-06 04:30:23 +0000 UTC" firstStartedPulling="2025-12-06 04:30:24.221484742 +0000 UTC m=+1413.227189903" lastFinishedPulling="2025-12-06 04:30:32.900072924 +0000 UTC m=+1421.905778125" observedRunningTime="2025-12-06 04:30:33.890826817 +0000 UTC m=+1422.896532038" watchObservedRunningTime="2025-12-06 04:30:33.894167437 +0000 UTC m=+1422.899872628" Dec 06 04:30:51 crc kubenswrapper[4718]: I1206 04:30:51.976132 4718 scope.go:117] "RemoveContainer" containerID="d4097c8b997fb1210a6abf71344ec95d75701d2c96a10b2355a22ac2cb4cfb58" Dec 06 04:30:52 crc kubenswrapper[4718]: I1206 04:30:51.998845 4718 scope.go:117] "RemoveContainer" containerID="c4e319dd489da8e7cb93d75fc7df1f3be46c078698a0dc8e8c695fe1c20bcd42" Dec 06 04:30:56 crc kubenswrapper[4718]: I1206 04:30:56.809476 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-czt9h"] Dec 06 04:30:56 crc kubenswrapper[4718]: I1206 04:30:56.812585 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:56 crc kubenswrapper[4718]: I1206 04:30:56.827750 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-czt9h"] Dec 06 04:30:56 crc kubenswrapper[4718]: I1206 04:30:56.996634 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fphkj\" (UniqueName: \"kubernetes.io/projected/e581aa40-ffba-4532-b349-f610c4fa676a-kube-api-access-fphkj\") pod \"redhat-operators-czt9h\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:56 crc kubenswrapper[4718]: I1206 04:30:56.996693 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-utilities\") pod \"redhat-operators-czt9h\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:56 crc kubenswrapper[4718]: I1206 04:30:56.996756 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-catalog-content\") pod \"redhat-operators-czt9h\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.098464 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fphkj\" (UniqueName: \"kubernetes.io/projected/e581aa40-ffba-4532-b349-f610c4fa676a-kube-api-access-fphkj\") pod \"redhat-operators-czt9h\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.098517 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-utilities\") pod \"redhat-operators-czt9h\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.098561 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-catalog-content\") pod \"redhat-operators-czt9h\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.099097 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-catalog-content\") pod \"redhat-operators-czt9h\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.099169 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-utilities\") pod \"redhat-operators-czt9h\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.121735 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fphkj\" (UniqueName: \"kubernetes.io/projected/e581aa40-ffba-4532-b349-f610c4fa676a-kube-api-access-fphkj\") pod \"redhat-operators-czt9h\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.142948 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.567194 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-czt9h"] Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.877200 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:30:57 crc kubenswrapper[4718]: I1206 04:30:57.877561 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:30:58 crc kubenswrapper[4718]: I1206 04:30:58.060710 4718 generic.go:334] "Generic (PLEG): container finished" podID="e581aa40-ffba-4532-b349-f610c4fa676a" containerID="1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0" exitCode=0 Dec 06 04:30:58 crc kubenswrapper[4718]: I1206 04:30:58.060767 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czt9h" event={"ID":"e581aa40-ffba-4532-b349-f610c4fa676a","Type":"ContainerDied","Data":"1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0"} Dec 06 04:30:58 crc kubenswrapper[4718]: I1206 04:30:58.060799 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czt9h" event={"ID":"e581aa40-ffba-4532-b349-f610c4fa676a","Type":"ContainerStarted","Data":"083f035f28744be2e92a5fbf27225db6b97bbf9ba0b73c94c8c0a77ad675e1d5"} Dec 06 04:31:00 crc kubenswrapper[4718]: I1206 04:31:00.080479 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czt9h" event={"ID":"e581aa40-ffba-4532-b349-f610c4fa676a","Type":"ContainerStarted","Data":"ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4"} Dec 06 04:31:01 crc kubenswrapper[4718]: I1206 04:31:01.089795 4718 generic.go:334] "Generic (PLEG): container finished" podID="e581aa40-ffba-4532-b349-f610c4fa676a" containerID="ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4" exitCode=0 Dec 06 04:31:01 crc kubenswrapper[4718]: I1206 04:31:01.089845 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czt9h" event={"ID":"e581aa40-ffba-4532-b349-f610c4fa676a","Type":"ContainerDied","Data":"ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4"} Dec 06 04:31:02 crc kubenswrapper[4718]: I1206 04:31:02.101037 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czt9h" event={"ID":"e581aa40-ffba-4532-b349-f610c4fa676a","Type":"ContainerStarted","Data":"0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9"} Dec 06 04:31:03 crc kubenswrapper[4718]: I1206 04:31:03.127730 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-czt9h" podStartSLOduration=3.397536057 podStartE2EDuration="7.127704736s" podCreationTimestamp="2025-12-06 04:30:56 +0000 UTC" firstStartedPulling="2025-12-06 04:30:58.062043089 +0000 UTC m=+1447.067748250" lastFinishedPulling="2025-12-06 04:31:01.792211758 +0000 UTC m=+1450.797916929" observedRunningTime="2025-12-06 04:31:03.125211909 +0000 UTC m=+1452.130917090" watchObservedRunningTime="2025-12-06 04:31:03.127704736 +0000 UTC m=+1452.133409897" Dec 06 04:31:07 crc kubenswrapper[4718]: I1206 04:31:07.143663 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:31:07 crc kubenswrapper[4718]: I1206 04:31:07.144013 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:31:08 crc kubenswrapper[4718]: I1206 04:31:08.200182 4718 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-czt9h" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" containerName="registry-server" probeResult="failure" output=< Dec 06 04:31:08 crc kubenswrapper[4718]: timeout: failed to connect service ":50051" within 1s Dec 06 04:31:08 crc kubenswrapper[4718]: > Dec 06 04:31:17 crc kubenswrapper[4718]: I1206 04:31:17.193886 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:31:17 crc kubenswrapper[4718]: I1206 04:31:17.246103 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:31:17 crc kubenswrapper[4718]: I1206 04:31:17.433700 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-czt9h"] Dec 06 04:31:18 crc kubenswrapper[4718]: I1206 04:31:18.238569 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-czt9h" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" containerName="registry-server" containerID="cri-o://0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9" gracePeriod=2 Dec 06 04:31:18 crc kubenswrapper[4718]: I1206 04:31:18.715032 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:31:18 crc kubenswrapper[4718]: I1206 04:31:18.857323 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fphkj\" (UniqueName: \"kubernetes.io/projected/e581aa40-ffba-4532-b349-f610c4fa676a-kube-api-access-fphkj\") pod \"e581aa40-ffba-4532-b349-f610c4fa676a\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " Dec 06 04:31:18 crc kubenswrapper[4718]: I1206 04:31:18.857394 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-utilities\") pod \"e581aa40-ffba-4532-b349-f610c4fa676a\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " Dec 06 04:31:18 crc kubenswrapper[4718]: I1206 04:31:18.857505 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-catalog-content\") pod \"e581aa40-ffba-4532-b349-f610c4fa676a\" (UID: \"e581aa40-ffba-4532-b349-f610c4fa676a\") " Dec 06 04:31:18 crc kubenswrapper[4718]: I1206 04:31:18.859368 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-utilities" (OuterVolumeSpecName: "utilities") pod "e581aa40-ffba-4532-b349-f610c4fa676a" (UID: "e581aa40-ffba-4532-b349-f610c4fa676a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:31:18 crc kubenswrapper[4718]: I1206 04:31:18.871694 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e581aa40-ffba-4532-b349-f610c4fa676a-kube-api-access-fphkj" (OuterVolumeSpecName: "kube-api-access-fphkj") pod "e581aa40-ffba-4532-b349-f610c4fa676a" (UID: "e581aa40-ffba-4532-b349-f610c4fa676a"). InnerVolumeSpecName "kube-api-access-fphkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:31:18 crc kubenswrapper[4718]: I1206 04:31:18.959607 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fphkj\" (UniqueName: \"kubernetes.io/projected/e581aa40-ffba-4532-b349-f610c4fa676a-kube-api-access-fphkj\") on node \"crc\" DevicePath \"\"" Dec 06 04:31:18 crc kubenswrapper[4718]: I1206 04:31:18.959659 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.032521 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e581aa40-ffba-4532-b349-f610c4fa676a" (UID: "e581aa40-ffba-4532-b349-f610c4fa676a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.061915 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e581aa40-ffba-4532-b349-f610c4fa676a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.252649 4718 generic.go:334] "Generic (PLEG): container finished" podID="e581aa40-ffba-4532-b349-f610c4fa676a" containerID="0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9" exitCode=0 Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.252709 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czt9h" event={"ID":"e581aa40-ffba-4532-b349-f610c4fa676a","Type":"ContainerDied","Data":"0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9"} Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.252732 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czt9h" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.252749 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czt9h" event={"ID":"e581aa40-ffba-4532-b349-f610c4fa676a","Type":"ContainerDied","Data":"083f035f28744be2e92a5fbf27225db6b97bbf9ba0b73c94c8c0a77ad675e1d5"} Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.252780 4718 scope.go:117] "RemoveContainer" containerID="0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.293494 4718 scope.go:117] "RemoveContainer" containerID="ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.311725 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-czt9h"] Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.319263 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-czt9h"] Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.330213 4718 scope.go:117] "RemoveContainer" containerID="1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.341749 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" path="/var/lib/kubelet/pods/e581aa40-ffba-4532-b349-f610c4fa676a/volumes" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.365905 4718 scope.go:117] "RemoveContainer" containerID="0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9" Dec 06 04:31:19 crc kubenswrapper[4718]: E1206 04:31:19.366436 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9\": container with ID starting with 0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9 not found: ID does not exist" containerID="0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.366478 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9"} err="failed to get container status \"0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9\": rpc error: code = NotFound desc = could not find container \"0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9\": container with ID starting with 0b9174275f4bcb88b0095c1ffc051071f1bdc64140c9b0de84ab485656af98b9 not found: ID does not exist" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.366503 4718 scope.go:117] "RemoveContainer" containerID="ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4" Dec 06 04:31:19 crc kubenswrapper[4718]: E1206 04:31:19.366844 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4\": container with ID starting with ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4 not found: ID does not exist" containerID="ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.366905 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4"} err="failed to get container status \"ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4\": rpc error: code = NotFound desc = could not find container \"ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4\": container with ID starting with ca20684bef6276fdff41d1056ba155545ce6545d63ff8a4134acfa88aee86fe4 not found: ID does not exist" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.366935 4718 scope.go:117] "RemoveContainer" containerID="1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0" Dec 06 04:31:19 crc kubenswrapper[4718]: E1206 04:31:19.367479 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0\": container with ID starting with 1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0 not found: ID does not exist" containerID="1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0" Dec 06 04:31:19 crc kubenswrapper[4718]: I1206 04:31:19.367501 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0"} err="failed to get container status \"1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0\": rpc error: code = NotFound desc = could not find container \"1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0\": container with ID starting with 1108370704bfc082864c16c0f94e72e4744601e4aee77f54ca9a70393507c7a0 not found: ID does not exist" Dec 06 04:31:27 crc kubenswrapper[4718]: I1206 04:31:27.877372 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:31:27 crc kubenswrapper[4718]: I1206 04:31:27.878183 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:31:27 crc kubenswrapper[4718]: I1206 04:31:27.878299 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:31:27 crc kubenswrapper[4718]: I1206 04:31:27.879362 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d81f022389d14fa68d0d6add32cf0d12240d0d061eea1e86215bd5dbdd1355c2"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:31:27 crc kubenswrapper[4718]: I1206 04:31:27.879489 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://d81f022389d14fa68d0d6add32cf0d12240d0d061eea1e86215bd5dbdd1355c2" gracePeriod=600 Dec 06 04:31:28 crc kubenswrapper[4718]: I1206 04:31:28.360367 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="d81f022389d14fa68d0d6add32cf0d12240d0d061eea1e86215bd5dbdd1355c2" exitCode=0 Dec 06 04:31:28 crc kubenswrapper[4718]: I1206 04:31:28.360419 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"d81f022389d14fa68d0d6add32cf0d12240d0d061eea1e86215bd5dbdd1355c2"} Dec 06 04:31:28 crc kubenswrapper[4718]: I1206 04:31:28.361464 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937"} Dec 06 04:31:28 crc kubenswrapper[4718]: I1206 04:31:28.361513 4718 scope.go:117] "RemoveContainer" containerID="8d135c41c1c0b733890b2561029cffaa9956d80b356ad7dc0217957931de3b87" Dec 06 04:31:52 crc kubenswrapper[4718]: I1206 04:31:52.063789 4718 scope.go:117] "RemoveContainer" containerID="397359323d68cd5bbe95c4bd8af8bf3b69594806e2ac558f21e05cfe69cb8580" Dec 06 04:31:52 crc kubenswrapper[4718]: I1206 04:31:52.101423 4718 scope.go:117] "RemoveContainer" containerID="ba3dc75586f7a5125e1d161f027763267ee39b7cbd3255acad43fa1b1aa5dd93" Dec 06 04:31:52 crc kubenswrapper[4718]: I1206 04:31:52.144012 4718 scope.go:117] "RemoveContainer" containerID="a1da07bd695de999dc679e8aea7d091aa409047b1878a5b0980ba4be4bc848f7" Dec 06 04:32:52 crc kubenswrapper[4718]: I1206 04:32:52.251114 4718 scope.go:117] "RemoveContainer" containerID="154af6f90e3efe9d74b39726a5b859e7935e42a7499c10b27f7046c0e470d9f0" Dec 06 04:32:52 crc kubenswrapper[4718]: I1206 04:32:52.299336 4718 scope.go:117] "RemoveContainer" containerID="d9b39207b2f4dcb5f75c163d4bb0aca0dec7dcb0fee345cd7d2d1c7cf677674a" Dec 06 04:32:52 crc kubenswrapper[4718]: I1206 04:32:52.349320 4718 scope.go:117] "RemoveContainer" containerID="93b6d25531f6f3944886d0e3bc606d3691ad13d1f0d63b31160049fd25dc9470" Dec 06 04:32:52 crc kubenswrapper[4718]: I1206 04:32:52.381076 4718 scope.go:117] "RemoveContainer" containerID="210b717444b8d92e2296cad127fce92445392521e154df5f6ef865bc21984a4f" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.086575 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xgws9"] Dec 06 04:33:30 crc kubenswrapper[4718]: E1206 04:33:30.088223 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" containerName="registry-server" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.088276 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" containerName="registry-server" Dec 06 04:33:30 crc kubenswrapper[4718]: E1206 04:33:30.088296 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" containerName="extract-utilities" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.088308 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" containerName="extract-utilities" Dec 06 04:33:30 crc kubenswrapper[4718]: E1206 04:33:30.088334 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" containerName="extract-content" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.088344 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" containerName="extract-content" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.088486 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e581aa40-ffba-4532-b349-f610c4fa676a" containerName="registry-server" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.089679 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.112839 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgws9"] Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.177000 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-utilities\") pod \"community-operators-xgws9\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.177102 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-catalog-content\") pod \"community-operators-xgws9\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.177132 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbf87\" (UniqueName: \"kubernetes.io/projected/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-kube-api-access-hbf87\") pod \"community-operators-xgws9\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.278952 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-catalog-content\") pod \"community-operators-xgws9\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.279015 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbf87\" (UniqueName: \"kubernetes.io/projected/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-kube-api-access-hbf87\") pod \"community-operators-xgws9\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.279103 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-utilities\") pod \"community-operators-xgws9\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.279563 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-catalog-content\") pod \"community-operators-xgws9\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.279596 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-utilities\") pod \"community-operators-xgws9\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.299083 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbf87\" (UniqueName: \"kubernetes.io/projected/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-kube-api-access-hbf87\") pod \"community-operators-xgws9\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.451744 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:30 crc kubenswrapper[4718]: I1206 04:33:30.931268 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgws9"] Dec 06 04:33:31 crc kubenswrapper[4718]: I1206 04:33:31.457597 4718 generic.go:334] "Generic (PLEG): container finished" podID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerID="5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29" exitCode=0 Dec 06 04:33:31 crc kubenswrapper[4718]: I1206 04:33:31.457683 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgws9" event={"ID":"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d","Type":"ContainerDied","Data":"5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29"} Dec 06 04:33:31 crc kubenswrapper[4718]: I1206 04:33:31.457972 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgws9" event={"ID":"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d","Type":"ContainerStarted","Data":"7e81bb1bc56a95dba7d9edaa87dc15ed9d01e7ba77564c2bed1c8e47398809ce"} Dec 06 04:33:32 crc kubenswrapper[4718]: I1206 04:33:32.469522 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgws9" event={"ID":"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d","Type":"ContainerStarted","Data":"0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0"} Dec 06 04:33:33 crc kubenswrapper[4718]: I1206 04:33:33.479456 4718 generic.go:334] "Generic (PLEG): container finished" podID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerID="0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0" exitCode=0 Dec 06 04:33:33 crc kubenswrapper[4718]: I1206 04:33:33.479518 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgws9" event={"ID":"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d","Type":"ContainerDied","Data":"0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0"} Dec 06 04:33:34 crc kubenswrapper[4718]: I1206 04:33:34.498390 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgws9" event={"ID":"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d","Type":"ContainerStarted","Data":"9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3"} Dec 06 04:33:34 crc kubenswrapper[4718]: I1206 04:33:34.518365 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xgws9" podStartSLOduration=2.053465037 podStartE2EDuration="4.518346644s" podCreationTimestamp="2025-12-06 04:33:30 +0000 UTC" firstStartedPulling="2025-12-06 04:33:31.45913177 +0000 UTC m=+1600.464836931" lastFinishedPulling="2025-12-06 04:33:33.924013337 +0000 UTC m=+1602.929718538" observedRunningTime="2025-12-06 04:33:34.511921372 +0000 UTC m=+1603.517626553" watchObservedRunningTime="2025-12-06 04:33:34.518346644 +0000 UTC m=+1603.524051815" Dec 06 04:33:40 crc kubenswrapper[4718]: I1206 04:33:40.452028 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:40 crc kubenswrapper[4718]: I1206 04:33:40.452975 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:40 crc kubenswrapper[4718]: I1206 04:33:40.505814 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:40 crc kubenswrapper[4718]: I1206 04:33:40.608543 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:40 crc kubenswrapper[4718]: I1206 04:33:40.760263 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgws9"] Dec 06 04:33:42 crc kubenswrapper[4718]: I1206 04:33:42.578975 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xgws9" podUID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerName="registry-server" containerID="cri-o://9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3" gracePeriod=2 Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.482766 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.501437 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-utilities\") pod \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.501524 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbf87\" (UniqueName: \"kubernetes.io/projected/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-kube-api-access-hbf87\") pod \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.501556 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-catalog-content\") pod \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\" (UID: \"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d\") " Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.502690 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-utilities" (OuterVolumeSpecName: "utilities") pod "3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" (UID: "3c639a4d-2f80-46a3-b7b0-1b31e1ba861d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.503081 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.509593 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-kube-api-access-hbf87" (OuterVolumeSpecName: "kube-api-access-hbf87") pod "3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" (UID: "3c639a4d-2f80-46a3-b7b0-1b31e1ba861d"). InnerVolumeSpecName "kube-api-access-hbf87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.549748 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" (UID: "3c639a4d-2f80-46a3-b7b0-1b31e1ba861d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.589937 4718 generic.go:334] "Generic (PLEG): container finished" podID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerID="9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3" exitCode=0 Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.590009 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgws9" event={"ID":"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d","Type":"ContainerDied","Data":"9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3"} Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.590058 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgws9" event={"ID":"3c639a4d-2f80-46a3-b7b0-1b31e1ba861d","Type":"ContainerDied","Data":"7e81bb1bc56a95dba7d9edaa87dc15ed9d01e7ba77564c2bed1c8e47398809ce"} Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.590103 4718 scope.go:117] "RemoveContainer" containerID="9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.590019 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgws9" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.603938 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbf87\" (UniqueName: \"kubernetes.io/projected/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-kube-api-access-hbf87\") on node \"crc\" DevicePath \"\"" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.603964 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.619333 4718 scope.go:117] "RemoveContainer" containerID="0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.639587 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgws9"] Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.643794 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xgws9"] Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.648450 4718 scope.go:117] "RemoveContainer" containerID="5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.695487 4718 scope.go:117] "RemoveContainer" containerID="9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3" Dec 06 04:33:43 crc kubenswrapper[4718]: E1206 04:33:43.696223 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3\": container with ID starting with 9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3 not found: ID does not exist" containerID="9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.696454 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3"} err="failed to get container status \"9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3\": rpc error: code = NotFound desc = could not find container \"9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3\": container with ID starting with 9567136c0f0f26d098f56e64ed9675160f300c7a7925540503651f9e3ce4ade3 not found: ID does not exist" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.696595 4718 scope.go:117] "RemoveContainer" containerID="0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0" Dec 06 04:33:43 crc kubenswrapper[4718]: E1206 04:33:43.697441 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0\": container with ID starting with 0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0 not found: ID does not exist" containerID="0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.697484 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0"} err="failed to get container status \"0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0\": rpc error: code = NotFound desc = could not find container \"0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0\": container with ID starting with 0eb926a17437e8b89c0b549090cf92cf0f18e72979cd0b82d04112f17309ebd0 not found: ID does not exist" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.697516 4718 scope.go:117] "RemoveContainer" containerID="5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29" Dec 06 04:33:43 crc kubenswrapper[4718]: E1206 04:33:43.698008 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29\": container with ID starting with 5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29 not found: ID does not exist" containerID="5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29" Dec 06 04:33:43 crc kubenswrapper[4718]: I1206 04:33:43.698052 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29"} err="failed to get container status \"5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29\": rpc error: code = NotFound desc = could not find container \"5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29\": container with ID starting with 5cecba43952a27fdc465d6f6ba0fe8890b9fcda95009929e6531fbb687945c29 not found: ID does not exist" Dec 06 04:33:45 crc kubenswrapper[4718]: I1206 04:33:45.343788 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" path="/var/lib/kubelet/pods/3c639a4d-2f80-46a3-b7b0-1b31e1ba861d/volumes" Dec 06 04:33:52 crc kubenswrapper[4718]: I1206 04:33:52.496563 4718 scope.go:117] "RemoveContainer" containerID="b0099011249fdf367d16392d4329508d27b73d1cef02576982f565954ecb143b" Dec 06 04:33:52 crc kubenswrapper[4718]: I1206 04:33:52.531400 4718 scope.go:117] "RemoveContainer" containerID="fbb003bd23b84746ad4abcfe0224e9c2e1c994221ebe6aab3271ece17e9b322f" Dec 06 04:33:52 crc kubenswrapper[4718]: I1206 04:33:52.576088 4718 scope.go:117] "RemoveContainer" containerID="9f02663b460b91cef7ba90f1d492a0f28e806adc30369a70208dcb0f584f4390" Dec 06 04:33:52 crc kubenswrapper[4718]: I1206 04:33:52.604759 4718 scope.go:117] "RemoveContainer" containerID="275894c18f169adb9c38bb81d34a65418eab583a1a05a8154ce5d60c2ff28dae" Dec 06 04:33:52 crc kubenswrapper[4718]: I1206 04:33:52.650215 4718 scope.go:117] "RemoveContainer" containerID="c1c5e05f6a64a1722daadfe2dfc587e9395af4b532db00f918ba84d596e9498a" Dec 06 04:33:57 crc kubenswrapper[4718]: I1206 04:33:57.878252 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:33:57 crc kubenswrapper[4718]: I1206 04:33:57.879782 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:34:17 crc kubenswrapper[4718]: I1206 04:34:17.907347 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qt52m"] Dec 06 04:34:17 crc kubenswrapper[4718]: E1206 04:34:17.907968 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerName="registry-server" Dec 06 04:34:17 crc kubenswrapper[4718]: I1206 04:34:17.908013 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerName="registry-server" Dec 06 04:34:17 crc kubenswrapper[4718]: E1206 04:34:17.908035 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerName="extract-content" Dec 06 04:34:17 crc kubenswrapper[4718]: I1206 04:34:17.908043 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerName="extract-content" Dec 06 04:34:17 crc kubenswrapper[4718]: E1206 04:34:17.908059 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerName="extract-utilities" Dec 06 04:34:17 crc kubenswrapper[4718]: I1206 04:34:17.908069 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerName="extract-utilities" Dec 06 04:34:17 crc kubenswrapper[4718]: I1206 04:34:17.908218 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c639a4d-2f80-46a3-b7b0-1b31e1ba861d" containerName="registry-server" Dec 06 04:34:17 crc kubenswrapper[4718]: I1206 04:34:17.909603 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:17 crc kubenswrapper[4718]: I1206 04:34:17.935289 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qt52m"] Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.024692 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-catalog-content\") pod \"redhat-marketplace-qt52m\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.024734 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-utilities\") pod \"redhat-marketplace-qt52m\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.024766 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5zhv\" (UniqueName: \"kubernetes.io/projected/bde18754-0b63-4278-b7ac-9720f5996f83-kube-api-access-n5zhv\") pod \"redhat-marketplace-qt52m\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.126101 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-catalog-content\") pod \"redhat-marketplace-qt52m\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.126178 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-utilities\") pod \"redhat-marketplace-qt52m\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.126263 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5zhv\" (UniqueName: \"kubernetes.io/projected/bde18754-0b63-4278-b7ac-9720f5996f83-kube-api-access-n5zhv\") pod \"redhat-marketplace-qt52m\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.126707 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-catalog-content\") pod \"redhat-marketplace-qt52m\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.126802 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-utilities\") pod \"redhat-marketplace-qt52m\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.146509 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5zhv\" (UniqueName: \"kubernetes.io/projected/bde18754-0b63-4278-b7ac-9720f5996f83-kube-api-access-n5zhv\") pod \"redhat-marketplace-qt52m\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.250161 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:18 crc kubenswrapper[4718]: I1206 04:34:18.480299 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qt52m"] Dec 06 04:34:19 crc kubenswrapper[4718]: I1206 04:34:19.083382 4718 generic.go:334] "Generic (PLEG): container finished" podID="bde18754-0b63-4278-b7ac-9720f5996f83" containerID="cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57" exitCode=0 Dec 06 04:34:19 crc kubenswrapper[4718]: I1206 04:34:19.083429 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qt52m" event={"ID":"bde18754-0b63-4278-b7ac-9720f5996f83","Type":"ContainerDied","Data":"cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57"} Dec 06 04:34:19 crc kubenswrapper[4718]: I1206 04:34:19.083460 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qt52m" event={"ID":"bde18754-0b63-4278-b7ac-9720f5996f83","Type":"ContainerStarted","Data":"ed2bd791b0bfedb6e6b274de50dcfea993ec21ce058dd7f0b45a06316d62d9b1"} Dec 06 04:34:20 crc kubenswrapper[4718]: I1206 04:34:20.093569 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qt52m" event={"ID":"bde18754-0b63-4278-b7ac-9720f5996f83","Type":"ContainerStarted","Data":"debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72"} Dec 06 04:34:21 crc kubenswrapper[4718]: I1206 04:34:21.105590 4718 generic.go:334] "Generic (PLEG): container finished" podID="bde18754-0b63-4278-b7ac-9720f5996f83" containerID="debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72" exitCode=0 Dec 06 04:34:21 crc kubenswrapper[4718]: I1206 04:34:21.105645 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qt52m" event={"ID":"bde18754-0b63-4278-b7ac-9720f5996f83","Type":"ContainerDied","Data":"debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72"} Dec 06 04:34:22 crc kubenswrapper[4718]: I1206 04:34:22.114848 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qt52m" event={"ID":"bde18754-0b63-4278-b7ac-9720f5996f83","Type":"ContainerStarted","Data":"bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547"} Dec 06 04:34:27 crc kubenswrapper[4718]: I1206 04:34:27.877212 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:34:27 crc kubenswrapper[4718]: I1206 04:34:27.878045 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:34:28 crc kubenswrapper[4718]: I1206 04:34:28.251097 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:28 crc kubenswrapper[4718]: I1206 04:34:28.252045 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:28 crc kubenswrapper[4718]: I1206 04:34:28.298829 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:28 crc kubenswrapper[4718]: I1206 04:34:28.322113 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qt52m" podStartSLOduration=8.909121359 podStartE2EDuration="11.322097274s" podCreationTimestamp="2025-12-06 04:34:17 +0000 UTC" firstStartedPulling="2025-12-06 04:34:19.085220613 +0000 UTC m=+1648.090925774" lastFinishedPulling="2025-12-06 04:34:21.498196528 +0000 UTC m=+1650.503901689" observedRunningTime="2025-12-06 04:34:22.14550039 +0000 UTC m=+1651.151205551" watchObservedRunningTime="2025-12-06 04:34:28.322097274 +0000 UTC m=+1657.327802435" Dec 06 04:34:29 crc kubenswrapper[4718]: I1206 04:34:29.219225 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:29 crc kubenswrapper[4718]: I1206 04:34:29.280207 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qt52m"] Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.194498 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qt52m" podUID="bde18754-0b63-4278-b7ac-9720f5996f83" containerName="registry-server" containerID="cri-o://bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547" gracePeriod=2 Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.678495 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.739294 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5zhv\" (UniqueName: \"kubernetes.io/projected/bde18754-0b63-4278-b7ac-9720f5996f83-kube-api-access-n5zhv\") pod \"bde18754-0b63-4278-b7ac-9720f5996f83\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.739503 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-catalog-content\") pod \"bde18754-0b63-4278-b7ac-9720f5996f83\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.739544 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-utilities\") pod \"bde18754-0b63-4278-b7ac-9720f5996f83\" (UID: \"bde18754-0b63-4278-b7ac-9720f5996f83\") " Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.740651 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-utilities" (OuterVolumeSpecName: "utilities") pod "bde18754-0b63-4278-b7ac-9720f5996f83" (UID: "bde18754-0b63-4278-b7ac-9720f5996f83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.744454 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde18754-0b63-4278-b7ac-9720f5996f83-kube-api-access-n5zhv" (OuterVolumeSpecName: "kube-api-access-n5zhv") pod "bde18754-0b63-4278-b7ac-9720f5996f83" (UID: "bde18754-0b63-4278-b7ac-9720f5996f83"). InnerVolumeSpecName "kube-api-access-n5zhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.764026 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bde18754-0b63-4278-b7ac-9720f5996f83" (UID: "bde18754-0b63-4278-b7ac-9720f5996f83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.841493 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.841531 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bde18754-0b63-4278-b7ac-9720f5996f83-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:34:31 crc kubenswrapper[4718]: I1206 04:34:31.841543 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5zhv\" (UniqueName: \"kubernetes.io/projected/bde18754-0b63-4278-b7ac-9720f5996f83-kube-api-access-n5zhv\") on node \"crc\" DevicePath \"\"" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.207596 4718 generic.go:334] "Generic (PLEG): container finished" podID="bde18754-0b63-4278-b7ac-9720f5996f83" containerID="bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547" exitCode=0 Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.207689 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qt52m" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.207687 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qt52m" event={"ID":"bde18754-0b63-4278-b7ac-9720f5996f83","Type":"ContainerDied","Data":"bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547"} Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.208107 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qt52m" event={"ID":"bde18754-0b63-4278-b7ac-9720f5996f83","Type":"ContainerDied","Data":"ed2bd791b0bfedb6e6b274de50dcfea993ec21ce058dd7f0b45a06316d62d9b1"} Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.208136 4718 scope.go:117] "RemoveContainer" containerID="bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.231139 4718 scope.go:117] "RemoveContainer" containerID="debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.261566 4718 scope.go:117] "RemoveContainer" containerID="cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.268403 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qt52m"] Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.283123 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qt52m"] Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.303010 4718 scope.go:117] "RemoveContainer" containerID="bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547" Dec 06 04:34:32 crc kubenswrapper[4718]: E1206 04:34:32.303836 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547\": container with ID starting with bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547 not found: ID does not exist" containerID="bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.303890 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547"} err="failed to get container status \"bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547\": rpc error: code = NotFound desc = could not find container \"bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547\": container with ID starting with bf42ed048149c260a8a4286d2ca9253f0df1d09d180e6dc41816492538661547 not found: ID does not exist" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.303923 4718 scope.go:117] "RemoveContainer" containerID="debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72" Dec 06 04:34:32 crc kubenswrapper[4718]: E1206 04:34:32.304477 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72\": container with ID starting with debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72 not found: ID does not exist" containerID="debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.304505 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72"} err="failed to get container status \"debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72\": rpc error: code = NotFound desc = could not find container \"debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72\": container with ID starting with debf97d903328c180de78e0eb0c7308b2a31c0ad8e7648522d37bb7ad17c5c72 not found: ID does not exist" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.304536 4718 scope.go:117] "RemoveContainer" containerID="cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57" Dec 06 04:34:32 crc kubenswrapper[4718]: E1206 04:34:32.304967 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57\": container with ID starting with cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57 not found: ID does not exist" containerID="cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57" Dec 06 04:34:32 crc kubenswrapper[4718]: I1206 04:34:32.305053 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57"} err="failed to get container status \"cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57\": rpc error: code = NotFound desc = could not find container \"cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57\": container with ID starting with cd33fb957734c1947bf314d1d69112f42c4cff227eb5d7b41865fff5985eac57 not found: ID does not exist" Dec 06 04:34:33 crc kubenswrapper[4718]: I1206 04:34:33.343186 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bde18754-0b63-4278-b7ac-9720f5996f83" path="/var/lib/kubelet/pods/bde18754-0b63-4278-b7ac-9720f5996f83/volumes" Dec 06 04:34:52 crc kubenswrapper[4718]: I1206 04:34:52.779047 4718 scope.go:117] "RemoveContainer" containerID="ecdc7a98dc8744e7eee8630449d7484f2f721680c7499e37a24d19601ae47bd5" Dec 06 04:34:52 crc kubenswrapper[4718]: I1206 04:34:52.810398 4718 scope.go:117] "RemoveContainer" containerID="a5c59c349ea3821949394f3f6698b50d0f399a9f755767d1d830b1b845c7e50f" Dec 06 04:34:52 crc kubenswrapper[4718]: I1206 04:34:52.851511 4718 scope.go:117] "RemoveContainer" containerID="bad8ba20b3c27557f13cd7a49404d358f44fc13fbe00db668711b3e7f2442a26" Dec 06 04:34:52 crc kubenswrapper[4718]: I1206 04:34:52.879300 4718 scope.go:117] "RemoveContainer" containerID="8692c0d26ab0ffc9e45c3d9d400c3b3b032bfa2cdd828a9e670dfa5810e3bdc2" Dec 06 04:34:52 crc kubenswrapper[4718]: I1206 04:34:52.914488 4718 scope.go:117] "RemoveContainer" containerID="2675050ca4fc9fafac602f52eaf368505d0e1afcc10b9eebe2914b5d5ab1858f" Dec 06 04:34:57 crc kubenswrapper[4718]: I1206 04:34:57.877279 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:34:57 crc kubenswrapper[4718]: I1206 04:34:57.877861 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:34:57 crc kubenswrapper[4718]: I1206 04:34:57.877910 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:34:57 crc kubenswrapper[4718]: I1206 04:34:57.878533 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:34:57 crc kubenswrapper[4718]: I1206 04:34:57.878608 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" gracePeriod=600 Dec 06 04:34:58 crc kubenswrapper[4718]: E1206 04:34:58.012185 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:34:58 crc kubenswrapper[4718]: I1206 04:34:58.452381 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" exitCode=0 Dec 06 04:34:58 crc kubenswrapper[4718]: I1206 04:34:58.452791 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937"} Dec 06 04:34:58 crc kubenswrapper[4718]: I1206 04:34:58.453112 4718 scope.go:117] "RemoveContainer" containerID="d81f022389d14fa68d0d6add32cf0d12240d0d061eea1e86215bd5dbdd1355c2" Dec 06 04:34:58 crc kubenswrapper[4718]: I1206 04:34:58.455750 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:34:58 crc kubenswrapper[4718]: E1206 04:34:58.456168 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:35:11 crc kubenswrapper[4718]: I1206 04:35:11.339516 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:35:11 crc kubenswrapper[4718]: E1206 04:35:11.340564 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:35:24 crc kubenswrapper[4718]: I1206 04:35:24.327797 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:35:24 crc kubenswrapper[4718]: E1206 04:35:24.328565 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:35:35 crc kubenswrapper[4718]: I1206 04:35:35.328283 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:35:35 crc kubenswrapper[4718]: E1206 04:35:35.329208 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:35:50 crc kubenswrapper[4718]: I1206 04:35:50.328497 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:35:50 crc kubenswrapper[4718]: E1206 04:35:50.329651 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:35:53 crc kubenswrapper[4718]: I1206 04:35:53.015638 4718 scope.go:117] "RemoveContainer" containerID="f89ac623a98973b93760a2f4df0bb78061ff34253647809e3821c0d65993aaba" Dec 06 04:36:03 crc kubenswrapper[4718]: I1206 04:36:03.328573 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:36:03 crc kubenswrapper[4718]: E1206 04:36:03.329462 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:36:16 crc kubenswrapper[4718]: I1206 04:36:16.329376 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:36:16 crc kubenswrapper[4718]: E1206 04:36:16.330494 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:36:28 crc kubenswrapper[4718]: I1206 04:36:28.329140 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:36:28 crc kubenswrapper[4718]: E1206 04:36:28.330520 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:36:42 crc kubenswrapper[4718]: I1206 04:36:42.328694 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:36:42 crc kubenswrapper[4718]: E1206 04:36:42.329489 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:36:56 crc kubenswrapper[4718]: I1206 04:36:56.327651 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:36:56 crc kubenswrapper[4718]: E1206 04:36:56.329614 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:37:10 crc kubenswrapper[4718]: I1206 04:37:10.328534 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:37:10 crc kubenswrapper[4718]: E1206 04:37:10.329533 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:37:24 crc kubenswrapper[4718]: I1206 04:37:24.328096 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:37:24 crc kubenswrapper[4718]: E1206 04:37:24.329069 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:37:37 crc kubenswrapper[4718]: I1206 04:37:37.330388 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:37:37 crc kubenswrapper[4718]: E1206 04:37:37.331355 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:37:49 crc kubenswrapper[4718]: I1206 04:37:49.328605 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:37:49 crc kubenswrapper[4718]: E1206 04:37:49.329780 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:38:00 crc kubenswrapper[4718]: I1206 04:38:00.329028 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:38:00 crc kubenswrapper[4718]: E1206 04:38:00.330068 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:38:14 crc kubenswrapper[4718]: I1206 04:38:14.328427 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:38:14 crc kubenswrapper[4718]: E1206 04:38:14.329666 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:38:25 crc kubenswrapper[4718]: I1206 04:38:25.329134 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:38:25 crc kubenswrapper[4718]: E1206 04:38:25.330198 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:38:39 crc kubenswrapper[4718]: I1206 04:38:39.328415 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:38:39 crc kubenswrapper[4718]: E1206 04:38:39.329459 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:38:52 crc kubenswrapper[4718]: I1206 04:38:52.329186 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:38:52 crc kubenswrapper[4718]: E1206 04:38:52.330219 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:39:04 crc kubenswrapper[4718]: I1206 04:39:04.327807 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:39:04 crc kubenswrapper[4718]: E1206 04:39:04.329374 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:39:16 crc kubenswrapper[4718]: I1206 04:39:16.327868 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:39:16 crc kubenswrapper[4718]: E1206 04:39:16.328767 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:39:28 crc kubenswrapper[4718]: I1206 04:39:28.328407 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:39:28 crc kubenswrapper[4718]: E1206 04:39:28.329150 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:39:38 crc kubenswrapper[4718]: I1206 04:39:38.056351 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-k9mww"] Dec 06 04:39:38 crc kubenswrapper[4718]: I1206 04:39:38.058664 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g"] Dec 06 04:39:38 crc kubenswrapper[4718]: I1206 04:39:38.062787 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bbee-account-create-update-lx49g"] Dec 06 04:39:38 crc kubenswrapper[4718]: I1206 04:39:38.067592 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-k9mww"] Dec 06 04:39:39 crc kubenswrapper[4718]: I1206 04:39:39.337811 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afa81b48-c76a-44ac-9770-94a17bc4e136" path="/var/lib/kubelet/pods/afa81b48-c76a-44ac-9770-94a17bc4e136/volumes" Dec 06 04:39:39 crc kubenswrapper[4718]: I1206 04:39:39.338955 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f13aaefa-a7d6-4ac6-b67f-1710f79b1791" path="/var/lib/kubelet/pods/f13aaefa-a7d6-4ac6-b67f-1710f79b1791/volumes" Dec 06 04:39:40 crc kubenswrapper[4718]: I1206 04:39:40.328451 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:39:40 crc kubenswrapper[4718]: E1206 04:39:40.328886 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:39:45 crc kubenswrapper[4718]: I1206 04:39:45.037508 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-zzlxc"] Dec 06 04:39:45 crc kubenswrapper[4718]: I1206 04:39:45.045801 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-zzlxc"] Dec 06 04:39:45 crc kubenswrapper[4718]: I1206 04:39:45.341219 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cf00cd1-3b58-4426-9d60-b5e16c2cafa2" path="/var/lib/kubelet/pods/5cf00cd1-3b58-4426-9d60-b5e16c2cafa2/volumes" Dec 06 04:39:51 crc kubenswrapper[4718]: I1206 04:39:51.040404 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-g7vxp"] Dec 06 04:39:51 crc kubenswrapper[4718]: I1206 04:39:51.049422 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-g7vxp"] Dec 06 04:39:51 crc kubenswrapper[4718]: I1206 04:39:51.339995 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e15091c-5c28-411a-906d-31e3860fcfe4" path="/var/lib/kubelet/pods/6e15091c-5c28-411a-906d-31e3860fcfe4/volumes" Dec 06 04:39:53 crc kubenswrapper[4718]: I1206 04:39:53.105308 4718 scope.go:117] "RemoveContainer" containerID="440203bc8764d63be0160e90ad48f81dee27949f692af3dc35136f88363a20e9" Dec 06 04:39:53 crc kubenswrapper[4718]: I1206 04:39:53.142519 4718 scope.go:117] "RemoveContainer" containerID="edb179856b3c24498b6199c969cc70cd97420e32028272d18b4c4779b076904f" Dec 06 04:39:53 crc kubenswrapper[4718]: I1206 04:39:53.184009 4718 scope.go:117] "RemoveContainer" containerID="1abd178aced5617cc967f88b78ccca1c13cb4b851d0855d6fe169e8c22de8027" Dec 06 04:39:53 crc kubenswrapper[4718]: I1206 04:39:53.203051 4718 scope.go:117] "RemoveContainer" containerID="5f29c1709190ab16ceb77a4fdb2742e157f5612881c1d659cc9bb610df21193a" Dec 06 04:39:55 crc kubenswrapper[4718]: I1206 04:39:55.328639 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:39:55 crc kubenswrapper[4718]: E1206 04:39:55.329557 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:40:07 crc kubenswrapper[4718]: I1206 04:40:07.332388 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:40:08 crc kubenswrapper[4718]: I1206 04:40:08.434790 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"ae94ee04e0f0b6f53a2fb12aac1d020fc0dc1260d9f11235b84486c47305e1c7"} Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.251091 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.251883 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/openstackclient" podUID="a7582aad-5aa4-494b-84c4-a164850e254f" containerName="openstackclient" containerID="cri-o://38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14" gracePeriod=30 Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.635394 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstackclient" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.665539 4718 generic.go:334] "Generic (PLEG): container finished" podID="a7582aad-5aa4-494b-84c4-a164850e254f" containerID="38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14" exitCode=143 Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.665575 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstackclient" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.665579 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstackclient" event={"ID":"a7582aad-5aa4-494b-84c4-a164850e254f","Type":"ContainerDied","Data":"38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14"} Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.665688 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstackclient" event={"ID":"a7582aad-5aa4-494b-84c4-a164850e254f","Type":"ContainerDied","Data":"6cf83db62fc8be6d4aaaaaf62e4f6e19468d4ca61247dd64a955829282cbad92"} Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.665705 4718 scope.go:117] "RemoveContainer" containerID="38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.683527 4718 scope.go:117] "RemoveContainer" containerID="38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14" Dec 06 04:40:39 crc kubenswrapper[4718]: E1206 04:40:39.684043 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14\": container with ID starting with 38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14 not found: ID does not exist" containerID="38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.684093 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14"} err="failed to get container status \"38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14\": rpc error: code = NotFound desc = could not find container \"38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14\": container with ID starting with 38901e1af6f3bce735bd06b15a888bb933fa84230981d7c4aa2ddc8dd8887b14 not found: ID does not exist" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.699800 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config-secret\") pod \"a7582aad-5aa4-494b-84c4-a164850e254f\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.699857 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config\") pod \"a7582aad-5aa4-494b-84c4-a164850e254f\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.699990 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpgs9\" (UniqueName: \"kubernetes.io/projected/a7582aad-5aa4-494b-84c4-a164850e254f-kube-api-access-xpgs9\") pod \"a7582aad-5aa4-494b-84c4-a164850e254f\" (UID: \"a7582aad-5aa4-494b-84c4-a164850e254f\") " Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.704975 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7582aad-5aa4-494b-84c4-a164850e254f-kube-api-access-xpgs9" (OuterVolumeSpecName: "kube-api-access-xpgs9") pod "a7582aad-5aa4-494b-84c4-a164850e254f" (UID: "a7582aad-5aa4-494b-84c4-a164850e254f"). InnerVolumeSpecName "kube-api-access-xpgs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.715194 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a7582aad-5aa4-494b-84c4-a164850e254f" (UID: "a7582aad-5aa4-494b-84c4-a164850e254f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.716563 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a7582aad-5aa4-494b-84c4-a164850e254f" (UID: "a7582aad-5aa4-494b-84c4-a164850e254f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.801774 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpgs9\" (UniqueName: \"kubernetes.io/projected/a7582aad-5aa4-494b-84c4-a164850e254f-kube-api-access-xpgs9\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.802139 4718 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:39 crc kubenswrapper[4718]: I1206 04:40:39.802223 4718 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7582aad-5aa4-494b-84c4-a164850e254f-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.000298 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.006634 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.184406 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2"] Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.184754 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" podUID="5549ac76-32a3-4ec8-a924-df653c1bf781" containerName="keystone-api" containerID="cri-o://8e632837865cd0d9e76c72b29ce8781b4510667e413fac0d212f2b8cdddf3f1e" gracePeriod=30 Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.229659 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystonebbee-account-delete-xhjs9"] Dec 06 04:40:40 crc kubenswrapper[4718]: E1206 04:40:40.230091 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde18754-0b63-4278-b7ac-9720f5996f83" containerName="registry-server" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.230188 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde18754-0b63-4278-b7ac-9720f5996f83" containerName="registry-server" Dec 06 04:40:40 crc kubenswrapper[4718]: E1206 04:40:40.230279 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde18754-0b63-4278-b7ac-9720f5996f83" containerName="extract-utilities" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.230346 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde18754-0b63-4278-b7ac-9720f5996f83" containerName="extract-utilities" Dec 06 04:40:40 crc kubenswrapper[4718]: E1206 04:40:40.230425 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7582aad-5aa4-494b-84c4-a164850e254f" containerName="openstackclient" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.230498 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7582aad-5aa4-494b-84c4-a164850e254f" containerName="openstackclient" Dec 06 04:40:40 crc kubenswrapper[4718]: E1206 04:40:40.230577 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde18754-0b63-4278-b7ac-9720f5996f83" containerName="extract-content" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.230644 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde18754-0b63-4278-b7ac-9720f5996f83" containerName="extract-content" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.230838 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7582aad-5aa4-494b-84c4-a164850e254f" containerName="openstackclient" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.230905 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde18754-0b63-4278-b7ac-9720f5996f83" containerName="registry-server" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.231448 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.244776 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonebbee-account-delete-xhjs9"] Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.309321 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccf5275-9394-44f3-9b45-fab230a916f3-operator-scripts\") pod \"keystonebbee-account-delete-xhjs9\" (UID: \"6ccf5275-9394-44f3-9b45-fab230a916f3\") " pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.310115 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzg44\" (UniqueName: \"kubernetes.io/projected/6ccf5275-9394-44f3-9b45-fab230a916f3-kube-api-access-gzg44\") pod \"keystonebbee-account-delete-xhjs9\" (UID: \"6ccf5275-9394-44f3-9b45-fab230a916f3\") " pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.411979 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccf5275-9394-44f3-9b45-fab230a916f3-operator-scripts\") pod \"keystonebbee-account-delete-xhjs9\" (UID: \"6ccf5275-9394-44f3-9b45-fab230a916f3\") " pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.412413 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzg44\" (UniqueName: \"kubernetes.io/projected/6ccf5275-9394-44f3-9b45-fab230a916f3-kube-api-access-gzg44\") pod \"keystonebbee-account-delete-xhjs9\" (UID: \"6ccf5275-9394-44f3-9b45-fab230a916f3\") " pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.413141 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccf5275-9394-44f3-9b45-fab230a916f3-operator-scripts\") pod \"keystonebbee-account-delete-xhjs9\" (UID: \"6ccf5275-9394-44f3-9b45-fab230a916f3\") " pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.429161 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzg44\" (UniqueName: \"kubernetes.io/projected/6ccf5275-9394-44f3-9b45-fab230a916f3-kube-api-access-gzg44\") pod \"keystonebbee-account-delete-xhjs9\" (UID: \"6ccf5275-9394-44f3-9b45-fab230a916f3\") " pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.547512 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:40 crc kubenswrapper[4718]: I1206 04:40:40.932929 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonebbee-account-delete-xhjs9"] Dec 06 04:40:41 crc kubenswrapper[4718]: I1206 04:40:41.359568 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7582aad-5aa4-494b-84c4-a164850e254f" path="/var/lib/kubelet/pods/a7582aad-5aa4-494b-84c4-a164850e254f/volumes" Dec 06 04:40:41 crc kubenswrapper[4718]: E1206 04:40:41.382525 4718 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ccf5275_9394_44f3_9b45_fab230a916f3.slice/crio-conmon-5200401f5f35796459f3b9f14f820430c7d3ee6cb352636a4f282c2ef4bba2a4.scope\": RecentStats: unable to find data in memory cache]" Dec 06 04:40:41 crc kubenswrapper[4718]: I1206 04:40:41.686587 4718 generic.go:334] "Generic (PLEG): container finished" podID="6ccf5275-9394-44f3-9b45-fab230a916f3" containerID="5200401f5f35796459f3b9f14f820430c7d3ee6cb352636a4f282c2ef4bba2a4" exitCode=0 Dec 06 04:40:41 crc kubenswrapper[4718]: I1206 04:40:41.686684 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" event={"ID":"6ccf5275-9394-44f3-9b45-fab230a916f3","Type":"ContainerDied","Data":"5200401f5f35796459f3b9f14f820430c7d3ee6cb352636a4f282c2ef4bba2a4"} Dec 06 04:40:41 crc kubenswrapper[4718]: I1206 04:40:41.686953 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" event={"ID":"6ccf5275-9394-44f3-9b45-fab230a916f3","Type":"ContainerStarted","Data":"2cbdfbb2a78742a7c048b15fbc93efb8afed59b1049681c87aef35b2eca3682f"} Dec 06 04:40:42 crc kubenswrapper[4718]: I1206 04:40:42.933112 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.051535 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccf5275-9394-44f3-9b45-fab230a916f3-operator-scripts\") pod \"6ccf5275-9394-44f3-9b45-fab230a916f3\" (UID: \"6ccf5275-9394-44f3-9b45-fab230a916f3\") " Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.051594 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzg44\" (UniqueName: \"kubernetes.io/projected/6ccf5275-9394-44f3-9b45-fab230a916f3-kube-api-access-gzg44\") pod \"6ccf5275-9394-44f3-9b45-fab230a916f3\" (UID: \"6ccf5275-9394-44f3-9b45-fab230a916f3\") " Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.052105 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ccf5275-9394-44f3-9b45-fab230a916f3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6ccf5275-9394-44f3-9b45-fab230a916f3" (UID: "6ccf5275-9394-44f3-9b45-fab230a916f3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.057113 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ccf5275-9394-44f3-9b45-fab230a916f3-kube-api-access-gzg44" (OuterVolumeSpecName: "kube-api-access-gzg44") pod "6ccf5275-9394-44f3-9b45-fab230a916f3" (UID: "6ccf5275-9394-44f3-9b45-fab230a916f3"). InnerVolumeSpecName "kube-api-access-gzg44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.153001 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccf5275-9394-44f3-9b45-fab230a916f3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.153044 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzg44\" (UniqueName: \"kubernetes.io/projected/6ccf5275-9394-44f3-9b45-fab230a916f3-kube-api-access-gzg44\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.703503 4718 generic.go:334] "Generic (PLEG): container finished" podID="5549ac76-32a3-4ec8-a924-df653c1bf781" containerID="8e632837865cd0d9e76c72b29ce8781b4510667e413fac0d212f2b8cdddf3f1e" exitCode=0 Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.703558 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" event={"ID":"5549ac76-32a3-4ec8-a924-df653c1bf781","Type":"ContainerDied","Data":"8e632837865cd0d9e76c72b29ce8781b4510667e413fac0d212f2b8cdddf3f1e"} Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.705248 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" event={"ID":"6ccf5275-9394-44f3-9b45-fab230a916f3","Type":"ContainerDied","Data":"2cbdfbb2a78742a7c048b15fbc93efb8afed59b1049681c87aef35b2eca3682f"} Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.705272 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cbdfbb2a78742a7c048b15fbc93efb8afed59b1049681c87aef35b2eca3682f" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.705277 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonebbee-account-delete-xhjs9" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.784260 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.925530 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-scripts\") pod \"5549ac76-32a3-4ec8-a924-df653c1bf781\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.925638 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-credential-keys\") pod \"5549ac76-32a3-4ec8-a924-df653c1bf781\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.925719 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52b8d\" (UniqueName: \"kubernetes.io/projected/5549ac76-32a3-4ec8-a924-df653c1bf781-kube-api-access-52b8d\") pod \"5549ac76-32a3-4ec8-a924-df653c1bf781\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.925840 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-fernet-keys\") pod \"5549ac76-32a3-4ec8-a924-df653c1bf781\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.925906 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-config-data\") pod \"5549ac76-32a3-4ec8-a924-df653c1bf781\" (UID: \"5549ac76-32a3-4ec8-a924-df653c1bf781\") " Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.929902 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-scripts" (OuterVolumeSpecName: "scripts") pod "5549ac76-32a3-4ec8-a924-df653c1bf781" (UID: "5549ac76-32a3-4ec8-a924-df653c1bf781"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.930604 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5549ac76-32a3-4ec8-a924-df653c1bf781-kube-api-access-52b8d" (OuterVolumeSpecName: "kube-api-access-52b8d") pod "5549ac76-32a3-4ec8-a924-df653c1bf781" (UID: "5549ac76-32a3-4ec8-a924-df653c1bf781"). InnerVolumeSpecName "kube-api-access-52b8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.931756 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5549ac76-32a3-4ec8-a924-df653c1bf781" (UID: "5549ac76-32a3-4ec8-a924-df653c1bf781"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.933584 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5549ac76-32a3-4ec8-a924-df653c1bf781" (UID: "5549ac76-32a3-4ec8-a924-df653c1bf781"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:43 crc kubenswrapper[4718]: I1206 04:40:43.951036 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-config-data" (OuterVolumeSpecName: "config-data") pod "5549ac76-32a3-4ec8-a924-df653c1bf781" (UID: "5549ac76-32a3-4ec8-a924-df653c1bf781"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.028246 4718 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.028284 4718 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.028299 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52b8d\" (UniqueName: \"kubernetes.io/projected/5549ac76-32a3-4ec8-a924-df653c1bf781-kube-api-access-52b8d\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.028309 4718 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.028325 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5549ac76-32a3-4ec8-a924-df653c1bf781-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.713458 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" event={"ID":"5549ac76-32a3-4ec8-a924-df653c1bf781","Type":"ContainerDied","Data":"18b0e1b59693d54b8a6eadd4dd8473af52b2d54111de149ced1216e82c0fc535"} Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.713534 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2" Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.713540 4718 scope.go:117] "RemoveContainer" containerID="8e632837865cd0d9e76c72b29ce8781b4510667e413fac0d212f2b8cdddf3f1e" Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.748885 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2"] Dec 06 04:40:44 crc kubenswrapper[4718]: I1206 04:40:44.757538 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-855fbcf9b4-f8gv2"] Dec 06 04:40:45 crc kubenswrapper[4718]: I1206 04:40:45.263891 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystonebbee-account-delete-xhjs9"] Dec 06 04:40:45 crc kubenswrapper[4718]: I1206 04:40:45.269794 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystonebbee-account-delete-xhjs9"] Dec 06 04:40:45 crc kubenswrapper[4718]: I1206 04:40:45.338655 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5549ac76-32a3-4ec8-a924-df653c1bf781" path="/var/lib/kubelet/pods/5549ac76-32a3-4ec8-a924-df653c1bf781/volumes" Dec 06 04:40:45 crc kubenswrapper[4718]: I1206 04:40:45.339629 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ccf5275-9394-44f3-9b45-fab230a916f3" path="/var/lib/kubelet/pods/6ccf5275-9394-44f3-9b45-fab230a916f3/volumes" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.765010 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vsb95"] Dec 06 04:40:50 crc kubenswrapper[4718]: E1206 04:40:50.765886 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5549ac76-32a3-4ec8-a924-df653c1bf781" containerName="keystone-api" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.765909 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="5549ac76-32a3-4ec8-a924-df653c1bf781" containerName="keystone-api" Dec 06 04:40:50 crc kubenswrapper[4718]: E1206 04:40:50.765935 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ccf5275-9394-44f3-9b45-fab230a916f3" containerName="mariadb-account-delete" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.765948 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ccf5275-9394-44f3-9b45-fab230a916f3" containerName="mariadb-account-delete" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.766182 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="5549ac76-32a3-4ec8-a924-df653c1bf781" containerName="keystone-api" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.766202 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ccf5275-9394-44f3-9b45-fab230a916f3" containerName="mariadb-account-delete" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.767733 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.793554 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vsb95"] Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.844631 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-catalog-content\") pod \"certified-operators-vsb95\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.844755 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxknr\" (UniqueName: \"kubernetes.io/projected/57f6d523-1a2c-43dd-a8f0-6589baccbc49-kube-api-access-dxknr\") pod \"certified-operators-vsb95\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.845046 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-utilities\") pod \"certified-operators-vsb95\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.946617 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-catalog-content\") pod \"certified-operators-vsb95\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.946672 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxknr\" (UniqueName: \"kubernetes.io/projected/57f6d523-1a2c-43dd-a8f0-6589baccbc49-kube-api-access-dxknr\") pod \"certified-operators-vsb95\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.946709 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-utilities\") pod \"certified-operators-vsb95\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.947684 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-catalog-content\") pod \"certified-operators-vsb95\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.947752 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-utilities\") pod \"certified-operators-vsb95\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:50 crc kubenswrapper[4718]: I1206 04:40:50.972290 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxknr\" (UniqueName: \"kubernetes.io/projected/57f6d523-1a2c-43dd-a8f0-6589baccbc49-kube-api-access-dxknr\") pod \"certified-operators-vsb95\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:51 crc kubenswrapper[4718]: I1206 04:40:51.092537 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:40:51 crc kubenswrapper[4718]: I1206 04:40:51.365046 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vsb95"] Dec 06 04:40:51 crc kubenswrapper[4718]: I1206 04:40:51.770724 4718 generic.go:334] "Generic (PLEG): container finished" podID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerID="f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de" exitCode=0 Dec 06 04:40:51 crc kubenswrapper[4718]: I1206 04:40:51.770792 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsb95" event={"ID":"57f6d523-1a2c-43dd-a8f0-6589baccbc49","Type":"ContainerDied","Data":"f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de"} Dec 06 04:40:51 crc kubenswrapper[4718]: I1206 04:40:51.771032 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsb95" event={"ID":"57f6d523-1a2c-43dd-a8f0-6589baccbc49","Type":"ContainerStarted","Data":"89f6453760c23e5ea3f58a8f8492fd4d2e2355fcb4e94f1f60b9d0e68cc6bac6"} Dec 06 04:40:51 crc kubenswrapper[4718]: I1206 04:40:51.772909 4718 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:40:52 crc kubenswrapper[4718]: I1206 04:40:52.455455 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Dec 06 04:40:52 crc kubenswrapper[4718]: I1206 04:40:52.464423 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Dec 06 04:40:52 crc kubenswrapper[4718]: I1206 04:40:52.469458 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Dec 06 04:40:52 crc kubenswrapper[4718]: I1206 04:40:52.598890 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/openstack-galera-2" podUID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" containerName="galera" containerID="cri-o://a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b" gracePeriod=30 Dec 06 04:40:52 crc kubenswrapper[4718]: I1206 04:40:52.778838 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsb95" event={"ID":"57f6d523-1a2c-43dd-a8f0-6589baccbc49","Type":"ContainerStarted","Data":"a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4"} Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.075602 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.075818 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/memcached-0" podUID="f1e994e7-dee9-487f-aab6-aa711e080dcc" containerName="memcached" containerID="cri-o://bdea561e2e9b28034932b4af606975c07a361078d7406c942c644d036717d304" gracePeriod=30 Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.460578 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.489716 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.570315 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.587331 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kolla-config\") pod \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.587385 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-operator-scripts\") pod \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.587454 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-generated\") pod \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.587490 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v7v5\" (UniqueName: \"kubernetes.io/projected/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kube-api-access-9v7v5\") pod \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.587513 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.587612 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-default\") pod \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\" (UID: \"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0\") " Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.587943 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" (UID: "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.588408 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" (UID: "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.588440 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" (UID: "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.588754 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" (UID: "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.594322 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kube-api-access-9v7v5" (OuterVolumeSpecName: "kube-api-access-9v7v5") pod "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" (UID: "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0"). InnerVolumeSpecName "kube-api-access-9v7v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.598472 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "mysql-db") pod "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" (UID: "e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.689045 4718 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.689318 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v7v5\" (UniqueName: \"kubernetes.io/projected/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kube-api-access-9v7v5\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.689416 4718 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.689487 4718 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.689567 4718 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.689639 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.705573 4718 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.786829 4718 generic.go:334] "Generic (PLEG): container finished" podID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" containerID="a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b" exitCode=0 Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.786901 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0","Type":"ContainerDied","Data":"a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b"} Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.786931 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0","Type":"ContainerDied","Data":"5624a19eb6e41bb9a6603f207f51818574de46d40a0a4e730741ddd42629babf"} Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.786952 4718 scope.go:117] "RemoveContainer" containerID="a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.787051 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-2" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.790207 4718 generic.go:334] "Generic (PLEG): container finished" podID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerID="a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4" exitCode=0 Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.790272 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsb95" event={"ID":"57f6d523-1a2c-43dd-a8f0-6589baccbc49","Type":"ContainerDied","Data":"a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4"} Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.790651 4718 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.807593 4718 scope.go:117] "RemoveContainer" containerID="6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.832884 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.833162 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/rabbitmq-server-0" podUID="e9868af3-c211-4dde-9701-669e579f484b" containerName="rabbitmq" containerID="cri-o://c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096" gracePeriod=604800 Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.845437 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.846894 4718 scope.go:117] "RemoveContainer" containerID="a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b" Dec 06 04:40:53 crc kubenswrapper[4718]: E1206 04:40:53.847648 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b\": container with ID starting with a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b not found: ID does not exist" containerID="a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.847697 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b"} err="failed to get container status \"a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b\": rpc error: code = NotFound desc = could not find container \"a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b\": container with ID starting with a7d59ab336d8db93ec33ce6d682beb1b54279ba190990613824e5f4173e4e79b not found: ID does not exist" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.847728 4718 scope.go:117] "RemoveContainer" containerID="6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6" Dec 06 04:40:53 crc kubenswrapper[4718]: E1206 04:40:53.848060 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6\": container with ID starting with 6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6 not found: ID does not exist" containerID="6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6" Dec 06 04:40:53 crc kubenswrapper[4718]: I1206 04:40:53.848093 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6"} err="failed to get container status \"6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6\": rpc error: code = NotFound desc = could not find container \"6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6\": container with ID starting with 6a86184559be7a7bb85e0d548e23ef1ca6251d2faa3dd496e75c3a95734292b6 not found: ID does not exist" Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.054293 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="keystone-kuttl-tests/memcached-0" podUID="f1e994e7-dee9-487f-aab6-aa711e080dcc" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.57:11211: connect: connection refused" Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.479882 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd"] Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.480112 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" podUID="8277a3e3-eae1-486b-936d-e921d8991932" containerName="manager" containerID="cri-o://6e61a0ec9d7b0d5ca50f9ae775dd6feea92c57bdd50ee7297f67732d5544e518" gracePeriod=10 Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.628217 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/openstack-galera-1" podUID="dde27b21-eedb-46d4-9cf5-99508535250e" containerName="galera" containerID="cri-o://bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285" gracePeriod=28 Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.732639 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-hql9d"] Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.732891 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-hql9d" podUID="147bc60c-8979-462d-bfc5-8460b289738d" containerName="registry-server" containerID="cri-o://d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4" gracePeriod=30 Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.774095 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh"] Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.777116 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/8f388e23ecf40c8e2d959df859cb17de619676ef573cf153d8c1e63851vssgh"] Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.798183 4718 generic.go:334] "Generic (PLEG): container finished" podID="f1e994e7-dee9-487f-aab6-aa711e080dcc" containerID="bdea561e2e9b28034932b4af606975c07a361078d7406c942c644d036717d304" exitCode=0 Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.798292 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/memcached-0" event={"ID":"f1e994e7-dee9-487f-aab6-aa711e080dcc","Type":"ContainerDied","Data":"bdea561e2e9b28034932b4af606975c07a361078d7406c942c644d036717d304"} Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.803719 4718 generic.go:334] "Generic (PLEG): container finished" podID="8277a3e3-eae1-486b-936d-e921d8991932" containerID="6e61a0ec9d7b0d5ca50f9ae775dd6feea92c57bdd50ee7297f67732d5544e518" exitCode=0 Dec 06 04:40:54 crc kubenswrapper[4718]: I1206 04:40:54.803757 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" event={"ID":"8277a3e3-eae1-486b-936d-e921d8991932","Type":"ContainerDied","Data":"6e61a0ec9d7b0d5ca50f9ae775dd6feea92c57bdd50ee7297f67732d5544e518"} Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.185676 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.210821 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq5w4\" (UniqueName: \"kubernetes.io/projected/147bc60c-8979-462d-bfc5-8460b289738d-kube-api-access-vq5w4\") pod \"147bc60c-8979-462d-bfc5-8460b289738d\" (UID: \"147bc60c-8979-462d-bfc5-8460b289738d\") " Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.219100 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147bc60c-8979-462d-bfc5-8460b289738d-kube-api-access-vq5w4" (OuterVolumeSpecName: "kube-api-access-vq5w4") pod "147bc60c-8979-462d-bfc5-8460b289738d" (UID: "147bc60c-8979-462d-bfc5-8460b289738d"). InnerVolumeSpecName "kube-api-access-vq5w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.312826 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq5w4\" (UniqueName: \"kubernetes.io/projected/147bc60c-8979-462d-bfc5-8460b289738d-kube-api-access-vq5w4\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.337116 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="092d522a-d186-4cfc-8f7e-1a337b4463ba" path="/var/lib/kubelet/pods/092d522a-d186-4cfc-8f7e-1a337b4463ba/volumes" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.337986 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" path="/var/lib/kubelet/pods/e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0/volumes" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.553654 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.638270 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/memcached-0" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.730725 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-apiservice-cert\") pod \"8277a3e3-eae1-486b-936d-e921d8991932\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.730781 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqp9z\" (UniqueName: \"kubernetes.io/projected/8277a3e3-eae1-486b-936d-e921d8991932-kube-api-access-bqp9z\") pod \"8277a3e3-eae1-486b-936d-e921d8991932\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.731226 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-webhook-cert\") pod \"8277a3e3-eae1-486b-936d-e921d8991932\" (UID: \"8277a3e3-eae1-486b-936d-e921d8991932\") " Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.737122 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "8277a3e3-eae1-486b-936d-e921d8991932" (UID: "8277a3e3-eae1-486b-936d-e921d8991932"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.737172 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "8277a3e3-eae1-486b-936d-e921d8991932" (UID: "8277a3e3-eae1-486b-936d-e921d8991932"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.737218 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8277a3e3-eae1-486b-936d-e921d8991932-kube-api-access-bqp9z" (OuterVolumeSpecName: "kube-api-access-bqp9z") pod "8277a3e3-eae1-486b-936d-e921d8991932" (UID: "8277a3e3-eae1-486b-936d-e921d8991932"). InnerVolumeSpecName "kube-api-access-bqp9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.811642 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" event={"ID":"8277a3e3-eae1-486b-936d-e921d8991932","Type":"ContainerDied","Data":"9e2414b1993cb4a69e32586595c37081528356472d3280e20571b8734c9eb71d"} Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.811699 4718 scope.go:117] "RemoveContainer" containerID="6e61a0ec9d7b0d5ca50f9ae775dd6feea92c57bdd50ee7297f67732d5544e518" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.811651 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.816696 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsb95" event={"ID":"57f6d523-1a2c-43dd-a8f0-6589baccbc49","Type":"ContainerStarted","Data":"dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3"} Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.820549 4718 generic.go:334] "Generic (PLEG): container finished" podID="147bc60c-8979-462d-bfc5-8460b289738d" containerID="d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4" exitCode=0 Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.820618 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hql9d" event={"ID":"147bc60c-8979-462d-bfc5-8460b289738d","Type":"ContainerDied","Data":"d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4"} Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.820646 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hql9d" event={"ID":"147bc60c-8979-462d-bfc5-8460b289738d","Type":"ContainerDied","Data":"b232968272c927c011084e0038f6ac71a0bedb88a94078a8c726e017065e5c98"} Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.820698 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hql9d" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.826196 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/memcached-0" event={"ID":"f1e994e7-dee9-487f-aab6-aa711e080dcc","Type":"ContainerDied","Data":"16c680d1112eb85b4abea4a597a860cb92fb4ad16e4e09e99840facfc45f4234"} Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.826288 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/memcached-0" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.829312 4718 scope.go:117] "RemoveContainer" containerID="d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.832745 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-kolla-config\") pod \"f1e994e7-dee9-487f-aab6-aa711e080dcc\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.832783 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f78dn\" (UniqueName: \"kubernetes.io/projected/f1e994e7-dee9-487f-aab6-aa711e080dcc-kube-api-access-f78dn\") pod \"f1e994e7-dee9-487f-aab6-aa711e080dcc\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.832838 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-config-data\") pod \"f1e994e7-dee9-487f-aab6-aa711e080dcc\" (UID: \"f1e994e7-dee9-487f-aab6-aa711e080dcc\") " Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.833078 4718 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.833089 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqp9z\" (UniqueName: \"kubernetes.io/projected/8277a3e3-eae1-486b-936d-e921d8991932-kube-api-access-bqp9z\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.833101 4718 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8277a3e3-eae1-486b-936d-e921d8991932-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.833901 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-config-data" (OuterVolumeSpecName: "config-data") pod "f1e994e7-dee9-487f-aab6-aa711e080dcc" (UID: "f1e994e7-dee9-487f-aab6-aa711e080dcc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.834506 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "f1e994e7-dee9-487f-aab6-aa711e080dcc" (UID: "f1e994e7-dee9-487f-aab6-aa711e080dcc"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.835780 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vsb95" podStartSLOduration=2.806573234 podStartE2EDuration="5.83576662s" podCreationTimestamp="2025-12-06 04:40:50 +0000 UTC" firstStartedPulling="2025-12-06 04:40:51.772656053 +0000 UTC m=+2040.778361214" lastFinishedPulling="2025-12-06 04:40:54.801849449 +0000 UTC m=+2043.807554600" observedRunningTime="2025-12-06 04:40:55.834159448 +0000 UTC m=+2044.839864609" watchObservedRunningTime="2025-12-06 04:40:55.83576662 +0000 UTC m=+2044.841471781" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.838010 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e994e7-dee9-487f-aab6-aa711e080dcc-kube-api-access-f78dn" (OuterVolumeSpecName: "kube-api-access-f78dn") pod "f1e994e7-dee9-487f-aab6-aa711e080dcc" (UID: "f1e994e7-dee9-487f-aab6-aa711e080dcc"). InnerVolumeSpecName "kube-api-access-f78dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.851929 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd"] Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.858004 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-94f9cff5f-65dbd"] Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.869893 4718 scope.go:117] "RemoveContainer" containerID="d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4" Dec 06 04:40:55 crc kubenswrapper[4718]: E1206 04:40:55.870434 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4\": container with ID starting with d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4 not found: ID does not exist" containerID="d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.870492 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4"} err="failed to get container status \"d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4\": rpc error: code = NotFound desc = could not find container \"d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4\": container with ID starting with d9b6bb984a4ba5f75bb86a2667862778f3d2623ddeba4fabc8f411865633e7b4 not found: ID does not exist" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.870523 4718 scope.go:117] "RemoveContainer" containerID="bdea561e2e9b28034932b4af606975c07a361078d7406c942c644d036717d304" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.884146 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-hql9d"] Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.897880 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-hql9d"] Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.934533 4718 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.934565 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f78dn\" (UniqueName: \"kubernetes.io/projected/f1e994e7-dee9-487f-aab6-aa711e080dcc-kube-api-access-f78dn\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:55 crc kubenswrapper[4718]: I1206 04:40:55.934575 4718 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1e994e7-dee9-487f-aab6-aa711e080dcc-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.212509 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.216734 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.277935 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.441936 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e9868af3-c211-4dde-9701-669e579f484b-erlang-cookie-secret\") pod \"e9868af3-c211-4dde-9701-669e579f484b\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.442044 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfcng\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-kube-api-access-hfcng\") pod \"e9868af3-c211-4dde-9701-669e579f484b\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.442091 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e9868af3-c211-4dde-9701-669e579f484b-pod-info\") pod \"e9868af3-c211-4dde-9701-669e579f484b\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.442151 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-plugins\") pod \"e9868af3-c211-4dde-9701-669e579f484b\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.442224 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e9868af3-c211-4dde-9701-669e579f484b-plugins-conf\") pod \"e9868af3-c211-4dde-9701-669e579f484b\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.442378 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\") pod \"e9868af3-c211-4dde-9701-669e579f484b\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.442434 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-confd\") pod \"e9868af3-c211-4dde-9701-669e579f484b\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.442499 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-erlang-cookie\") pod \"e9868af3-c211-4dde-9701-669e579f484b\" (UID: \"e9868af3-c211-4dde-9701-669e579f484b\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.443201 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e9868af3-c211-4dde-9701-669e579f484b" (UID: "e9868af3-c211-4dde-9701-669e579f484b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.443964 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9868af3-c211-4dde-9701-669e579f484b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e9868af3-c211-4dde-9701-669e579f484b" (UID: "e9868af3-c211-4dde-9701-669e579f484b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.444455 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e9868af3-c211-4dde-9701-669e579f484b" (UID: "e9868af3-c211-4dde-9701-669e579f484b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.446437 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e9868af3-c211-4dde-9701-669e579f484b-pod-info" (OuterVolumeSpecName: "pod-info") pod "e9868af3-c211-4dde-9701-669e579f484b" (UID: "e9868af3-c211-4dde-9701-669e579f484b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.446681 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-kube-api-access-hfcng" (OuterVolumeSpecName: "kube-api-access-hfcng") pod "e9868af3-c211-4dde-9701-669e579f484b" (UID: "e9868af3-c211-4dde-9701-669e579f484b"). InnerVolumeSpecName "kube-api-access-hfcng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.447074 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9868af3-c211-4dde-9701-669e579f484b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e9868af3-c211-4dde-9701-669e579f484b" (UID: "e9868af3-c211-4dde-9701-669e579f484b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.452939 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59" (OuterVolumeSpecName: "persistence") pod "e9868af3-c211-4dde-9701-669e579f484b" (UID: "e9868af3-c211-4dde-9701-669e579f484b"). InnerVolumeSpecName "pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.524210 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e9868af3-c211-4dde-9701-669e579f484b" (UID: "e9868af3-c211-4dde-9701-669e579f484b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.543877 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfcng\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-kube-api-access-hfcng\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.543911 4718 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e9868af3-c211-4dde-9701-669e579f484b-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.543920 4718 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.543929 4718 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e9868af3-c211-4dde-9701-669e579f484b-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.543959 4718 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\") on node \"crc\" " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.543970 4718 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.543979 4718 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e9868af3-c211-4dde-9701-669e579f484b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.543990 4718 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e9868af3-c211-4dde-9701-669e579f484b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.555193 4718 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.555358 4718 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59") on node "crc" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.573202 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.645125 4718 reconciler_common.go:293] "Volume detached for volume \"pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2839268f-929b-4394-9bc2-a8fcc4de8e59\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.652747 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/openstack-galera-0" podUID="225ba523-35eb-409e-8b11-9d1ec8994b38" containerName="galera" containerID="cri-o://49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282" gracePeriod=26 Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.745822 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-kolla-config\") pod \"dde27b21-eedb-46d4-9cf5-99508535250e\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.745961 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-default\") pod \"dde27b21-eedb-46d4-9cf5-99508535250e\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.746042 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"dde27b21-eedb-46d4-9cf5-99508535250e\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.746103 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmkc6\" (UniqueName: \"kubernetes.io/projected/dde27b21-eedb-46d4-9cf5-99508535250e-kube-api-access-gmkc6\") pod \"dde27b21-eedb-46d4-9cf5-99508535250e\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.746346 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-generated\") pod \"dde27b21-eedb-46d4-9cf5-99508535250e\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.746405 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-operator-scripts\") pod \"dde27b21-eedb-46d4-9cf5-99508535250e\" (UID: \"dde27b21-eedb-46d4-9cf5-99508535250e\") " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.746714 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "dde27b21-eedb-46d4-9cf5-99508535250e" (UID: "dde27b21-eedb-46d4-9cf5-99508535250e"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.746775 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "dde27b21-eedb-46d4-9cf5-99508535250e" (UID: "dde27b21-eedb-46d4-9cf5-99508535250e"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.747037 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "dde27b21-eedb-46d4-9cf5-99508535250e" (UID: "dde27b21-eedb-46d4-9cf5-99508535250e"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.747330 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dde27b21-eedb-46d4-9cf5-99508535250e" (UID: "dde27b21-eedb-46d4-9cf5-99508535250e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.751568 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dde27b21-eedb-46d4-9cf5-99508535250e-kube-api-access-gmkc6" (OuterVolumeSpecName: "kube-api-access-gmkc6") pod "dde27b21-eedb-46d4-9cf5-99508535250e" (UID: "dde27b21-eedb-46d4-9cf5-99508535250e"). InnerVolumeSpecName "kube-api-access-gmkc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.756500 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "mysql-db") pod "dde27b21-eedb-46d4-9cf5-99508535250e" (UID: "dde27b21-eedb-46d4-9cf5-99508535250e"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.835635 4718 generic.go:334] "Generic (PLEG): container finished" podID="dde27b21-eedb-46d4-9cf5-99508535250e" containerID="bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285" exitCode=0 Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.835730 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"dde27b21-eedb-46d4-9cf5-99508535250e","Type":"ContainerDied","Data":"bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285"} Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.835761 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"dde27b21-eedb-46d4-9cf5-99508535250e","Type":"ContainerDied","Data":"ffa83ff9075b47cfa5bc94b654cf3190a65a8f3525ecf1ff25736b77e5a79a0f"} Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.835754 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-1" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.835817 4718 scope.go:117] "RemoveContainer" containerID="bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.849814 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmkc6\" (UniqueName: \"kubernetes.io/projected/dde27b21-eedb-46d4-9cf5-99508535250e-kube-api-access-gmkc6\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.849845 4718 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.849858 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.849870 4718 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.849884 4718 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/dde27b21-eedb-46d4-9cf5-99508535250e-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.849920 4718 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.853058 4718 generic.go:334] "Generic (PLEG): container finished" podID="e9868af3-c211-4dde-9701-669e579f484b" containerID="c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096" exitCode=0 Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.853144 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/rabbitmq-server-0" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.853186 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"e9868af3-c211-4dde-9701-669e579f484b","Type":"ContainerDied","Data":"c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096"} Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.853220 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"e9868af3-c211-4dde-9701-669e579f484b","Type":"ContainerDied","Data":"339e2b442af7dbb894f958c84f1b16d4f17090248a4696daf5c7f88eb4332f14"} Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.863662 4718 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.871104 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.873300 4718 scope.go:117] "RemoveContainer" containerID="246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.896841 4718 scope.go:117] "RemoveContainer" containerID="bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285" Dec 06 04:40:56 crc kubenswrapper[4718]: E1206 04:40:56.897368 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285\": container with ID starting with bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285 not found: ID does not exist" containerID="bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.897432 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285"} err="failed to get container status \"bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285\": rpc error: code = NotFound desc = could not find container \"bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285\": container with ID starting with bcba876a841cc01ca76d4f3f04671dede29d60d2b1d628a13cd1ecd557a7d285 not found: ID does not exist" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.897461 4718 scope.go:117] "RemoveContainer" containerID="246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d" Dec 06 04:40:56 crc kubenswrapper[4718]: E1206 04:40:56.898636 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d\": container with ID starting with 246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d not found: ID does not exist" containerID="246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.898666 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d"} err="failed to get container status \"246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d\": rpc error: code = NotFound desc = could not find container \"246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d\": container with ID starting with 246c8b326a1c9e198072a3e3fc5fa0af5b9369f55d959b710cac3c6676a5d70d not found: ID does not exist" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.898685 4718 scope.go:117] "RemoveContainer" containerID="c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.905038 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.917675 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.921570 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.926808 4718 scope.go:117] "RemoveContainer" containerID="1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.950867 4718 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.955809 4718 scope.go:117] "RemoveContainer" containerID="c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096" Dec 06 04:40:56 crc kubenswrapper[4718]: E1206 04:40:56.956419 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096\": container with ID starting with c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096 not found: ID does not exist" containerID="c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.956500 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096"} err="failed to get container status \"c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096\": rpc error: code = NotFound desc = could not find container \"c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096\": container with ID starting with c5fa6e089c6abb8e6b3754beffe2aca2dde99b36a055f0ab19fb8af096cfe096 not found: ID does not exist" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.956530 4718 scope.go:117] "RemoveContainer" containerID="1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc" Dec 06 04:40:56 crc kubenswrapper[4718]: E1206 04:40:56.956923 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc\": container with ID starting with 1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc not found: ID does not exist" containerID="1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc" Dec 06 04:40:56 crc kubenswrapper[4718]: I1206 04:40:56.956952 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc"} err="failed to get container status \"1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc\": rpc error: code = NotFound desc = could not find container \"1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc\": container with ID starting with 1b2a688e03253128899b3aff14e362ca84e844e220195d09e2fee63788dfe3bc not found: ID does not exist" Dec 06 04:40:56 crc kubenswrapper[4718]: E1206 04:40:56.976432 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 04:40:56 crc kubenswrapper[4718]: E1206 04:40:56.978520 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 04:40:56 crc kubenswrapper[4718]: E1206 04:40:56.980180 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 06 04:40:56 crc kubenswrapper[4718]: E1206 04:40:56.980257 4718 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="keystone-kuttl-tests/openstack-galera-0" podUID="225ba523-35eb-409e-8b11-9d1ec8994b38" containerName="galera" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.334884 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147bc60c-8979-462d-bfc5-8460b289738d" path="/var/lib/kubelet/pods/147bc60c-8979-462d-bfc5-8460b289738d/volumes" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.335386 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8277a3e3-eae1-486b-936d-e921d8991932" path="/var/lib/kubelet/pods/8277a3e3-eae1-486b-936d-e921d8991932/volumes" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.335880 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dde27b21-eedb-46d4-9cf5-99508535250e" path="/var/lib/kubelet/pods/dde27b21-eedb-46d4-9cf5-99508535250e/volumes" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.337051 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9868af3-c211-4dde-9701-669e579f484b" path="/var/lib/kubelet/pods/e9868af3-c211-4dde-9701-669e579f484b/volumes" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.337598 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e994e7-dee9-487f-aab6-aa711e080dcc" path="/var/lib/kubelet/pods/f1e994e7-dee9-487f-aab6-aa711e080dcc/volumes" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.346466 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.393503 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc"] Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.393704 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" podUID="c51916d5-75f4-4435-b364-fc181ad7d21f" containerName="manager" containerID="cri-o://60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90" gracePeriod=10 Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.424305 4718 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" podUID="c51916d5-75f4-4435-b364-fc181ad7d21f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.53:8081/readyz\": dial tcp 10.217.0.53:8081: connect: connection refused" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.455916 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-generated\") pod \"225ba523-35eb-409e-8b11-9d1ec8994b38\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.455979 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-operator-scripts\") pod \"225ba523-35eb-409e-8b11-9d1ec8994b38\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.455998 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-kolla-config\") pod \"225ba523-35eb-409e-8b11-9d1ec8994b38\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.456040 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"225ba523-35eb-409e-8b11-9d1ec8994b38\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.456061 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-default\") pod \"225ba523-35eb-409e-8b11-9d1ec8994b38\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.456086 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tldhn\" (UniqueName: \"kubernetes.io/projected/225ba523-35eb-409e-8b11-9d1ec8994b38-kube-api-access-tldhn\") pod \"225ba523-35eb-409e-8b11-9d1ec8994b38\" (UID: \"225ba523-35eb-409e-8b11-9d1ec8994b38\") " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.456447 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "225ba523-35eb-409e-8b11-9d1ec8994b38" (UID: "225ba523-35eb-409e-8b11-9d1ec8994b38"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.456534 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "225ba523-35eb-409e-8b11-9d1ec8994b38" (UID: "225ba523-35eb-409e-8b11-9d1ec8994b38"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.456721 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "225ba523-35eb-409e-8b11-9d1ec8994b38" (UID: "225ba523-35eb-409e-8b11-9d1ec8994b38"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.457255 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "225ba523-35eb-409e-8b11-9d1ec8994b38" (UID: "225ba523-35eb-409e-8b11-9d1ec8994b38"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.460181 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/225ba523-35eb-409e-8b11-9d1ec8994b38-kube-api-access-tldhn" (OuterVolumeSpecName: "kube-api-access-tldhn") pod "225ba523-35eb-409e-8b11-9d1ec8994b38" (UID: "225ba523-35eb-409e-8b11-9d1ec8994b38"). InnerVolumeSpecName "kube-api-access-tldhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.466895 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "mysql-db") pod "225ba523-35eb-409e-8b11-9d1ec8994b38" (UID: "225ba523-35eb-409e-8b11-9d1ec8994b38"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.557860 4718 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.557895 4718 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.557906 4718 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.557940 4718 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.557952 4718 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/225ba523-35eb-409e-8b11-9d1ec8994b38-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.557961 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tldhn\" (UniqueName: \"kubernetes.io/projected/225ba523-35eb-409e-8b11-9d1ec8994b38-kube-api-access-tldhn\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.569210 4718 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.663038 4718 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.676671 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-8nhbw"] Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.676859 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-8nhbw" podUID="4ca99b94-f080-4ebc-9010-6aa0a0bcfac4" containerName="registry-server" containerID="cri-o://95deed81ab3e6f1a9ee4eca0594e53dbcbbc93fa0b1bdbb10f574c8c5df70273" gracePeriod=30 Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.690219 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9"] Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.695987 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/3392687e5dc44d06261c22d2d6f1c5816e47b94cbe8a8597be46f6f110ncgv9"] Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.827386 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.866675 4718 generic.go:334] "Generic (PLEG): container finished" podID="225ba523-35eb-409e-8b11-9d1ec8994b38" containerID="49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282" exitCode=0 Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.866730 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"225ba523-35eb-409e-8b11-9d1ec8994b38","Type":"ContainerDied","Data":"49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282"} Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.866755 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"225ba523-35eb-409e-8b11-9d1ec8994b38","Type":"ContainerDied","Data":"2d587eae184b515ac35c3adcec3b88de68dcdcb1480c91ac4bd7c43496f26dd1"} Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.866771 4718 scope.go:117] "RemoveContainer" containerID="49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.866844 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-0" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.873793 4718 generic.go:334] "Generic (PLEG): container finished" podID="c51916d5-75f4-4435-b364-fc181ad7d21f" containerID="60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90" exitCode=0 Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.873846 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" event={"ID":"c51916d5-75f4-4435-b364-fc181ad7d21f","Type":"ContainerDied","Data":"60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90"} Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.873865 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" event={"ID":"c51916d5-75f4-4435-b364-fc181ad7d21f","Type":"ContainerDied","Data":"db321dc8c9ada02785910a9602caf1e5bbceca1541d7ea6082d2dd721f571029"} Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.873908 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.882959 4718 generic.go:334] "Generic (PLEG): container finished" podID="4ca99b94-f080-4ebc-9010-6aa0a0bcfac4" containerID="95deed81ab3e6f1a9ee4eca0594e53dbcbbc93fa0b1bdbb10f574c8c5df70273" exitCode=0 Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.882994 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8nhbw" event={"ID":"4ca99b94-f080-4ebc-9010-6aa0a0bcfac4","Type":"ContainerDied","Data":"95deed81ab3e6f1a9ee4eca0594e53dbcbbc93fa0b1bdbb10f574c8c5df70273"} Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.889788 4718 scope.go:117] "RemoveContainer" containerID="ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.912380 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.918485 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.921623 4718 scope.go:117] "RemoveContainer" containerID="49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282" Dec 06 04:40:57 crc kubenswrapper[4718]: E1206 04:40:57.922611 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282\": container with ID starting with 49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282 not found: ID does not exist" containerID="49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.922653 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282"} err="failed to get container status \"49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282\": rpc error: code = NotFound desc = could not find container \"49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282\": container with ID starting with 49bf7f8335b9bc96175fef3b2b6e0b6886ad27d2406793eef0358c871c50f282 not found: ID does not exist" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.922672 4718 scope.go:117] "RemoveContainer" containerID="ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13" Dec 06 04:40:57 crc kubenswrapper[4718]: E1206 04:40:57.923548 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13\": container with ID starting with ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13 not found: ID does not exist" containerID="ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.923606 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13"} err="failed to get container status \"ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13\": rpc error: code = NotFound desc = could not find container \"ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13\": container with ID starting with ae84dfe9c71815c79037e5496959e99f9293d9a60da22c836107b9a081973a13 not found: ID does not exist" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.923634 4718 scope.go:117] "RemoveContainer" containerID="60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.948417 4718 scope.go:117] "RemoveContainer" containerID="60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90" Dec 06 04:40:57 crc kubenswrapper[4718]: E1206 04:40:57.950189 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90\": container with ID starting with 60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90 not found: ID does not exist" containerID="60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.950214 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90"} err="failed to get container status \"60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90\": rpc error: code = NotFound desc = could not find container \"60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90\": container with ID starting with 60298f53cc684a3146632b85718ac0b905dedfa25ba0b5d9715384aedfbc0a90 not found: ID does not exist" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.965958 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr5h8\" (UniqueName: \"kubernetes.io/projected/c51916d5-75f4-4435-b364-fc181ad7d21f-kube-api-access-mr5h8\") pod \"c51916d5-75f4-4435-b364-fc181ad7d21f\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.966048 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-apiservice-cert\") pod \"c51916d5-75f4-4435-b364-fc181ad7d21f\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.967193 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-webhook-cert\") pod \"c51916d5-75f4-4435-b364-fc181ad7d21f\" (UID: \"c51916d5-75f4-4435-b364-fc181ad7d21f\") " Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.970359 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51916d5-75f4-4435-b364-fc181ad7d21f-kube-api-access-mr5h8" (OuterVolumeSpecName: "kube-api-access-mr5h8") pod "c51916d5-75f4-4435-b364-fc181ad7d21f" (UID: "c51916d5-75f4-4435-b364-fc181ad7d21f"). InnerVolumeSpecName "kube-api-access-mr5h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.970390 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "c51916d5-75f4-4435-b364-fc181ad7d21f" (UID: "c51916d5-75f4-4435-b364-fc181ad7d21f"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:57 crc kubenswrapper[4718]: I1206 04:40:57.970501 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "c51916d5-75f4-4435-b364-fc181ad7d21f" (UID: "c51916d5-75f4-4435-b364-fc181ad7d21f"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.027586 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.068106 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkdcd\" (UniqueName: \"kubernetes.io/projected/4ca99b94-f080-4ebc-9010-6aa0a0bcfac4-kube-api-access-nkdcd\") pod \"4ca99b94-f080-4ebc-9010-6aa0a0bcfac4\" (UID: \"4ca99b94-f080-4ebc-9010-6aa0a0bcfac4\") " Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.068663 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr5h8\" (UniqueName: \"kubernetes.io/projected/c51916d5-75f4-4435-b364-fc181ad7d21f-kube-api-access-mr5h8\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.068700 4718 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.068718 4718 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c51916d5-75f4-4435-b364-fc181ad7d21f-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.071723 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca99b94-f080-4ebc-9010-6aa0a0bcfac4-kube-api-access-nkdcd" (OuterVolumeSpecName: "kube-api-access-nkdcd") pod "4ca99b94-f080-4ebc-9010-6aa0a0bcfac4" (UID: "4ca99b94-f080-4ebc-9010-6aa0a0bcfac4"). InnerVolumeSpecName "kube-api-access-nkdcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.169206 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkdcd\" (UniqueName: \"kubernetes.io/projected/4ca99b94-f080-4ebc-9010-6aa0a0bcfac4-kube-api-access-nkdcd\") on node \"crc\" DevicePath \"\"" Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.208592 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc"] Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.215322 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6df9d48c5f-fxqxc"] Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.891738 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8nhbw" event={"ID":"4ca99b94-f080-4ebc-9010-6aa0a0bcfac4","Type":"ContainerDied","Data":"4a57a2d515794ce33486d1240391e2e3f9fb0efb79a4e819906eb2ad788a231a"} Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.891782 4718 scope.go:117] "RemoveContainer" containerID="95deed81ab3e6f1a9ee4eca0594e53dbcbbc93fa0b1bdbb10f574c8c5df70273" Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.891811 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8nhbw" Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.922814 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-8nhbw"] Dec 06 04:40:58 crc kubenswrapper[4718]: I1206 04:40:58.926165 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-8nhbw"] Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.251942 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph"] Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.252204 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" podUID="043bf4f1-afde-4885-ab7a-764c626964c2" containerName="manager" containerID="cri-o://101fe35aa482af5d13bc0b558f367dad441240baacbb4b3557992a9703e4306c" gracePeriod=10 Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.338827 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="225ba523-35eb-409e-8b11-9d1ec8994b38" path="/var/lib/kubelet/pods/225ba523-35eb-409e-8b11-9d1ec8994b38/volumes" Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.339703 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ca99b94-f080-4ebc-9010-6aa0a0bcfac4" path="/var/lib/kubelet/pods/4ca99b94-f080-4ebc-9010-6aa0a0bcfac4/volumes" Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.340314 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="766a146c-c056-4843-be3b-fd5d4572624c" path="/var/lib/kubelet/pods/766a146c-c056-4843-be3b-fd5d4572624c/volumes" Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.341690 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c51916d5-75f4-4435-b364-fc181ad7d21f" path="/var/lib/kubelet/pods/c51916d5-75f4-4435-b364-fc181ad7d21f/volumes" Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.470592 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-zgxbt"] Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.470823 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-zgxbt" podUID="f6373f59-2255-4710-b1c2-e2d2c4624289" containerName="registry-server" containerID="cri-o://175b20f6fd66204f9be95f80d7dae08ed830a2f29f7eb7dccffed6ead1bd7c54" gracePeriod=30 Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.507215 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx"] Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.511831 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61v58lx"] Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.898252 4718 generic.go:334] "Generic (PLEG): container finished" podID="f6373f59-2255-4710-b1c2-e2d2c4624289" containerID="175b20f6fd66204f9be95f80d7dae08ed830a2f29f7eb7dccffed6ead1bd7c54" exitCode=0 Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.898321 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zgxbt" event={"ID":"f6373f59-2255-4710-b1c2-e2d2c4624289","Type":"ContainerDied","Data":"175b20f6fd66204f9be95f80d7dae08ed830a2f29f7eb7dccffed6ead1bd7c54"} Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.899732 4718 generic.go:334] "Generic (PLEG): container finished" podID="043bf4f1-afde-4885-ab7a-764c626964c2" containerID="101fe35aa482af5d13bc0b558f367dad441240baacbb4b3557992a9703e4306c" exitCode=0 Dec 06 04:40:59 crc kubenswrapper[4718]: I1206 04:40:59.899778 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" event={"ID":"043bf4f1-afde-4885-ab7a-764c626964c2","Type":"ContainerDied","Data":"101fe35aa482af5d13bc0b558f367dad441240baacbb4b3557992a9703e4306c"} Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.149904 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.295523 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w89n\" (UniqueName: \"kubernetes.io/projected/043bf4f1-afde-4885-ab7a-764c626964c2-kube-api-access-2w89n\") pod \"043bf4f1-afde-4885-ab7a-764c626964c2\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.295572 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-webhook-cert\") pod \"043bf4f1-afde-4885-ab7a-764c626964c2\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.295676 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-apiservice-cert\") pod \"043bf4f1-afde-4885-ab7a-764c626964c2\" (UID: \"043bf4f1-afde-4885-ab7a-764c626964c2\") " Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.299129 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "043bf4f1-afde-4885-ab7a-764c626964c2" (UID: "043bf4f1-afde-4885-ab7a-764c626964c2"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.299439 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "043bf4f1-afde-4885-ab7a-764c626964c2" (UID: "043bf4f1-afde-4885-ab7a-764c626964c2"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.299777 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/043bf4f1-afde-4885-ab7a-764c626964c2-kube-api-access-2w89n" (OuterVolumeSpecName: "kube-api-access-2w89n") pod "043bf4f1-afde-4885-ab7a-764c626964c2" (UID: "043bf4f1-afde-4885-ab7a-764c626964c2"). InnerVolumeSpecName "kube-api-access-2w89n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.322942 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.396599 4718 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.396637 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w89n\" (UniqueName: \"kubernetes.io/projected/043bf4f1-afde-4885-ab7a-764c626964c2-kube-api-access-2w89n\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.396687 4718 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/043bf4f1-afde-4885-ab7a-764c626964c2-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.497406 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9qwn\" (UniqueName: \"kubernetes.io/projected/f6373f59-2255-4710-b1c2-e2d2c4624289-kube-api-access-h9qwn\") pod \"f6373f59-2255-4710-b1c2-e2d2c4624289\" (UID: \"f6373f59-2255-4710-b1c2-e2d2c4624289\") " Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.501702 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6373f59-2255-4710-b1c2-e2d2c4624289-kube-api-access-h9qwn" (OuterVolumeSpecName: "kube-api-access-h9qwn") pod "f6373f59-2255-4710-b1c2-e2d2c4624289" (UID: "f6373f59-2255-4710-b1c2-e2d2c4624289"). InnerVolumeSpecName "kube-api-access-h9qwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.600404 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9qwn\" (UniqueName: \"kubernetes.io/projected/f6373f59-2255-4710-b1c2-e2d2c4624289-kube-api-access-h9qwn\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.909068 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zgxbt" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.909712 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zgxbt" event={"ID":"f6373f59-2255-4710-b1c2-e2d2c4624289","Type":"ContainerDied","Data":"8b2089920bb2df6a4b57fde423ed2015d92cd20bf81390326164bd61c9bec54c"} Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.910064 4718 scope.go:117] "RemoveContainer" containerID="175b20f6fd66204f9be95f80d7dae08ed830a2f29f7eb7dccffed6ead1bd7c54" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.911712 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" event={"ID":"043bf4f1-afde-4885-ab7a-764c626964c2","Type":"ContainerDied","Data":"39b2b1fd2ec436722773fa99932d4c6d2ac400ddbf935c56f4409652d529795b"} Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.911785 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.925865 4718 scope.go:117] "RemoveContainer" containerID="101fe35aa482af5d13bc0b558f367dad441240baacbb4b3557992a9703e4306c" Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.946048 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-zgxbt"] Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.951412 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-zgxbt"] Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.965443 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph"] Dec 06 04:41:00 crc kubenswrapper[4718]: I1206 04:41:00.970467 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-548f66d7bb-wn4ph"] Dec 06 04:41:01 crc kubenswrapper[4718]: I1206 04:41:01.093494 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:41:01 crc kubenswrapper[4718]: I1206 04:41:01.093565 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:41:01 crc kubenswrapper[4718]: I1206 04:41:01.132456 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:41:01 crc kubenswrapper[4718]: I1206 04:41:01.337105 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="043bf4f1-afde-4885-ab7a-764c626964c2" path="/var/lib/kubelet/pods/043bf4f1-afde-4885-ab7a-764c626964c2/volumes" Dec 06 04:41:01 crc kubenswrapper[4718]: I1206 04:41:01.338103 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7325a25-f02a-4afa-867f-98255311c258" path="/var/lib/kubelet/pods/e7325a25-f02a-4afa-867f-98255311c258/volumes" Dec 06 04:41:01 crc kubenswrapper[4718]: I1206 04:41:01.339214 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6373f59-2255-4710-b1c2-e2d2c4624289" path="/var/lib/kubelet/pods/f6373f59-2255-4710-b1c2-e2d2c4624289/volumes" Dec 06 04:41:01 crc kubenswrapper[4718]: I1206 04:41:01.954878 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.055960 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw"] Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.056191 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" podUID="513e6338-5c23-4b84-96b8-bd15003f6595" containerName="operator" containerID="cri-o://7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62" gracePeriod=10 Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.340370 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-9w2ld"] Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.340937 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" podUID="720bb2e9-e779-4b59-ac02-6f420b4dda74" containerName="registry-server" containerID="cri-o://31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3" gracePeriod=30 Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.371344 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb"] Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.375253 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590lfhwb"] Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.490621 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.625560 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpvrb\" (UniqueName: \"kubernetes.io/projected/513e6338-5c23-4b84-96b8-bd15003f6595-kube-api-access-vpvrb\") pod \"513e6338-5c23-4b84-96b8-bd15003f6595\" (UID: \"513e6338-5c23-4b84-96b8-bd15003f6595\") " Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.631841 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/513e6338-5c23-4b84-96b8-bd15003f6595-kube-api-access-vpvrb" (OuterVolumeSpecName: "kube-api-access-vpvrb") pod "513e6338-5c23-4b84-96b8-bd15003f6595" (UID: "513e6338-5c23-4b84-96b8-bd15003f6595"). InnerVolumeSpecName "kube-api-access-vpvrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.723994 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.726668 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpvrb\" (UniqueName: \"kubernetes.io/projected/513e6338-5c23-4b84-96b8-bd15003f6595-kube-api-access-vpvrb\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.827819 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ls96\" (UniqueName: \"kubernetes.io/projected/720bb2e9-e779-4b59-ac02-6f420b4dda74-kube-api-access-5ls96\") pod \"720bb2e9-e779-4b59-ac02-6f420b4dda74\" (UID: \"720bb2e9-e779-4b59-ac02-6f420b4dda74\") " Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.831572 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/720bb2e9-e779-4b59-ac02-6f420b4dda74-kube-api-access-5ls96" (OuterVolumeSpecName: "kube-api-access-5ls96") pod "720bb2e9-e779-4b59-ac02-6f420b4dda74" (UID: "720bb2e9-e779-4b59-ac02-6f420b4dda74"). InnerVolumeSpecName "kube-api-access-5ls96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.925732 4718 generic.go:334] "Generic (PLEG): container finished" podID="513e6338-5c23-4b84-96b8-bd15003f6595" containerID="7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62" exitCode=0 Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.925791 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.925796 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" event={"ID":"513e6338-5c23-4b84-96b8-bd15003f6595","Type":"ContainerDied","Data":"7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62"} Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.925956 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw" event={"ID":"513e6338-5c23-4b84-96b8-bd15003f6595","Type":"ContainerDied","Data":"f9e78616187c0b42d6d72501f4533a6633fdfce78667db7a8e145a5e88da3a3e"} Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.925999 4718 scope.go:117] "RemoveContainer" containerID="7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.927706 4718 generic.go:334] "Generic (PLEG): container finished" podID="720bb2e9-e779-4b59-ac02-6f420b4dda74" containerID="31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3" exitCode=0 Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.927734 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" event={"ID":"720bb2e9-e779-4b59-ac02-6f420b4dda74","Type":"ContainerDied","Data":"31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3"} Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.927772 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" event={"ID":"720bb2e9-e779-4b59-ac02-6f420b4dda74","Type":"ContainerDied","Data":"f466930d7eb850b6ae1403b1fd5b84bc9fac81b82edb38f70ae8148ce463d43f"} Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.927742 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-9w2ld" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.929495 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ls96\" (UniqueName: \"kubernetes.io/projected/720bb2e9-e779-4b59-ac02-6f420b4dda74-kube-api-access-5ls96\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.942907 4718 scope.go:117] "RemoveContainer" containerID="7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62" Dec 06 04:41:02 crc kubenswrapper[4718]: E1206 04:41:02.943205 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62\": container with ID starting with 7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62 not found: ID does not exist" containerID="7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.943260 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62"} err="failed to get container status \"7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62\": rpc error: code = NotFound desc = could not find container \"7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62\": container with ID starting with 7036f6d3b51b9b94787ac22ccf7f06d09bdab6d1e9fca74775f2f38b27294b62 not found: ID does not exist" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.943279 4718 scope.go:117] "RemoveContainer" containerID="31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.956209 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-9w2ld"] Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.964394 4718 scope.go:117] "RemoveContainer" containerID="31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.964927 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-9w2ld"] Dec 06 04:41:02 crc kubenswrapper[4718]: E1206 04:41:02.965489 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3\": container with ID starting with 31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3 not found: ID does not exist" containerID="31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.965520 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3"} err="failed to get container status \"31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3\": rpc error: code = NotFound desc = could not find container \"31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3\": container with ID starting with 31f88f52e43a740a9919dcc3a59c05f4517a97869ebc0513a104ca48e753c7d3 not found: ID does not exist" Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.966660 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw"] Dec 06 04:41:02 crc kubenswrapper[4718]: I1206 04:41:02.970426 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-cb7pw"] Dec 06 04:41:03 crc kubenswrapper[4718]: I1206 04:41:03.334285 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="513e6338-5c23-4b84-96b8-bd15003f6595" path="/var/lib/kubelet/pods/513e6338-5c23-4b84-96b8-bd15003f6595/volumes" Dec 06 04:41:03 crc kubenswrapper[4718]: I1206 04:41:03.334875 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f14df8d-3e33-475f-8e17-f507f1aaf429" path="/var/lib/kubelet/pods/5f14df8d-3e33-475f-8e17-f507f1aaf429/volumes" Dec 06 04:41:03 crc kubenswrapper[4718]: I1206 04:41:03.335522 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="720bb2e9-e779-4b59-ac02-6f420b4dda74" path="/var/lib/kubelet/pods/720bb2e9-e779-4b59-ac02-6f420b4dda74/volumes" Dec 06 04:41:03 crc kubenswrapper[4718]: I1206 04:41:03.746665 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vsb95"] Dec 06 04:41:03 crc kubenswrapper[4718]: I1206 04:41:03.935200 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vsb95" podUID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerName="registry-server" containerID="cri-o://dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3" gracePeriod=2 Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.779501 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.854380 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxknr\" (UniqueName: \"kubernetes.io/projected/57f6d523-1a2c-43dd-a8f0-6589baccbc49-kube-api-access-dxknr\") pod \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.854462 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-utilities\") pod \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.854503 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-catalog-content\") pod \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\" (UID: \"57f6d523-1a2c-43dd-a8f0-6589baccbc49\") " Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.855499 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-utilities" (OuterVolumeSpecName: "utilities") pod "57f6d523-1a2c-43dd-a8f0-6589baccbc49" (UID: "57f6d523-1a2c-43dd-a8f0-6589baccbc49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.862672 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57f6d523-1a2c-43dd-a8f0-6589baccbc49-kube-api-access-dxknr" (OuterVolumeSpecName: "kube-api-access-dxknr") pod "57f6d523-1a2c-43dd-a8f0-6589baccbc49" (UID: "57f6d523-1a2c-43dd-a8f0-6589baccbc49"). InnerVolumeSpecName "kube-api-access-dxknr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.919863 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57f6d523-1a2c-43dd-a8f0-6589baccbc49" (UID: "57f6d523-1a2c-43dd-a8f0-6589baccbc49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.942600 4718 generic.go:334] "Generic (PLEG): container finished" podID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerID="dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3" exitCode=0 Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.942649 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsb95" event={"ID":"57f6d523-1a2c-43dd-a8f0-6589baccbc49","Type":"ContainerDied","Data":"dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3"} Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.942678 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsb95" event={"ID":"57f6d523-1a2c-43dd-a8f0-6589baccbc49","Type":"ContainerDied","Data":"89f6453760c23e5ea3f58a8f8492fd4d2e2355fcb4e94f1f60b9d0e68cc6bac6"} Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.942697 4718 scope.go:117] "RemoveContainer" containerID="dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.942701 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsb95" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.955765 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxknr\" (UniqueName: \"kubernetes.io/projected/57f6d523-1a2c-43dd-a8f0-6589baccbc49-kube-api-access-dxknr\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.955798 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.955811 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57f6d523-1a2c-43dd-a8f0-6589baccbc49-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.958566 4718 scope.go:117] "RemoveContainer" containerID="a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4" Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.974695 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vsb95"] Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.978013 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vsb95"] Dec 06 04:41:04 crc kubenswrapper[4718]: I1206 04:41:04.994498 4718 scope.go:117] "RemoveContainer" containerID="f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de" Dec 06 04:41:05 crc kubenswrapper[4718]: I1206 04:41:05.015828 4718 scope.go:117] "RemoveContainer" containerID="dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3" Dec 06 04:41:05 crc kubenswrapper[4718]: E1206 04:41:05.016310 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3\": container with ID starting with dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3 not found: ID does not exist" containerID="dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3" Dec 06 04:41:05 crc kubenswrapper[4718]: I1206 04:41:05.016351 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3"} err="failed to get container status \"dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3\": rpc error: code = NotFound desc = could not find container \"dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3\": container with ID starting with dee15a8c8627a2dd1da4b7745374b9d92b88a89c21772795df18bdf038de23a3 not found: ID does not exist" Dec 06 04:41:05 crc kubenswrapper[4718]: I1206 04:41:05.016380 4718 scope.go:117] "RemoveContainer" containerID="a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4" Dec 06 04:41:05 crc kubenswrapper[4718]: E1206 04:41:05.016981 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4\": container with ID starting with a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4 not found: ID does not exist" containerID="a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4" Dec 06 04:41:05 crc kubenswrapper[4718]: I1206 04:41:05.017009 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4"} err="failed to get container status \"a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4\": rpc error: code = NotFound desc = could not find container \"a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4\": container with ID starting with a64d40aedd947329afaedc19b91638d6a8dbd2bdd77689d78ca067a1c52cfbd4 not found: ID does not exist" Dec 06 04:41:05 crc kubenswrapper[4718]: I1206 04:41:05.017026 4718 scope.go:117] "RemoveContainer" containerID="f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de" Dec 06 04:41:05 crc kubenswrapper[4718]: E1206 04:41:05.017281 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de\": container with ID starting with f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de not found: ID does not exist" containerID="f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de" Dec 06 04:41:05 crc kubenswrapper[4718]: I1206 04:41:05.017310 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de"} err="failed to get container status \"f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de\": rpc error: code = NotFound desc = could not find container \"f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de\": container with ID starting with f4428f06c2c1658c9fab779c9d4d5a5a430dd06df2add7a28dff1f73711f38de not found: ID does not exist" Dec 06 04:41:05 crc kubenswrapper[4718]: I1206 04:41:05.336035 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" path="/var/lib/kubelet/pods/57f6d523-1a2c-43dd-a8f0-6589baccbc49/volumes" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.359711 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xwscz"] Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360463 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147bc60c-8979-462d-bfc5-8460b289738d" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360490 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="147bc60c-8979-462d-bfc5-8460b289738d" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360511 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="513e6338-5c23-4b84-96b8-bd15003f6595" containerName="operator" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360524 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="513e6338-5c23-4b84-96b8-bd15003f6595" containerName="operator" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360547 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9868af3-c211-4dde-9701-669e579f484b" containerName="setup-container" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360561 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9868af3-c211-4dde-9701-669e579f484b" containerName="setup-container" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360579 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6373f59-2255-4710-b1c2-e2d2c4624289" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360592 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6373f59-2255-4710-b1c2-e2d2c4624289" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360608 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8277a3e3-eae1-486b-936d-e921d8991932" containerName="manager" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360621 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="8277a3e3-eae1-486b-936d-e921d8991932" containerName="manager" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360643 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerName="extract-utilities" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360655 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerName="extract-utilities" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360675 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="720bb2e9-e779-4b59-ac02-6f420b4dda74" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360687 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="720bb2e9-e779-4b59-ac02-6f420b4dda74" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360705 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9868af3-c211-4dde-9701-669e579f484b" containerName="rabbitmq" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360718 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9868af3-c211-4dde-9701-669e579f484b" containerName="rabbitmq" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360740 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51916d5-75f4-4435-b364-fc181ad7d21f" containerName="manager" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360752 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51916d5-75f4-4435-b364-fc181ad7d21f" containerName="manager" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360770 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e994e7-dee9-487f-aab6-aa711e080dcc" containerName="memcached" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360782 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e994e7-dee9-487f-aab6-aa711e080dcc" containerName="memcached" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360797 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerName="extract-content" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360809 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerName="extract-content" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360829 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225ba523-35eb-409e-8b11-9d1ec8994b38" containerName="galera" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360843 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="225ba523-35eb-409e-8b11-9d1ec8994b38" containerName="galera" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360863 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="043bf4f1-afde-4885-ab7a-764c626964c2" containerName="manager" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360880 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="043bf4f1-afde-4885-ab7a-764c626964c2" containerName="manager" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360907 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" containerName="galera" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360923 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" containerName="galera" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360942 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde27b21-eedb-46d4-9cf5-99508535250e" containerName="mysql-bootstrap" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.360955 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde27b21-eedb-46d4-9cf5-99508535250e" containerName="mysql-bootstrap" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.360974 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde27b21-eedb-46d4-9cf5-99508535250e" containerName="galera" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361072 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde27b21-eedb-46d4-9cf5-99508535250e" containerName="galera" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.361099 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225ba523-35eb-409e-8b11-9d1ec8994b38" containerName="mysql-bootstrap" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361137 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="225ba523-35eb-409e-8b11-9d1ec8994b38" containerName="mysql-bootstrap" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.361161 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" containerName="mysql-bootstrap" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361173 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" containerName="mysql-bootstrap" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.361202 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361267 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: E1206 04:41:08.361297 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca99b94-f080-4ebc-9010-6aa0a0bcfac4" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361313 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca99b94-f080-4ebc-9010-6aa0a0bcfac4" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361547 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="720bb2e9-e779-4b59-ac02-6f420b4dda74" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361571 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9868af3-c211-4dde-9701-669e579f484b" containerName="rabbitmq" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361587 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="225ba523-35eb-409e-8b11-9d1ec8994b38" containerName="galera" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361621 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="513e6338-5c23-4b84-96b8-bd15003f6595" containerName="operator" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361642 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca99b94-f080-4ebc-9010-6aa0a0bcfac4" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361662 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6373f59-2255-4710-b1c2-e2d2c4624289" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361686 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="8277a3e3-eae1-486b-936d-e921d8991932" containerName="manager" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361703 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="147bc60c-8979-462d-bfc5-8460b289738d" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361718 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="e102e7c8-0fc5-4b77-80ab-f1c3d37dcac0" containerName="galera" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361735 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e994e7-dee9-487f-aab6-aa711e080dcc" containerName="memcached" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361750 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="57f6d523-1a2c-43dd-a8f0-6589baccbc49" containerName="registry-server" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361770 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="dde27b21-eedb-46d4-9cf5-99508535250e" containerName="galera" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361791 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="043bf4f1-afde-4885-ab7a-764c626964c2" containerName="manager" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.361812 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51916d5-75f4-4435-b364-fc181ad7d21f" containerName="manager" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.363642 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.425723 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwscz"] Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.499914 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69hjq\" (UniqueName: \"kubernetes.io/projected/f120d8e7-a318-4238-ab2d-022b06d89ab7-kube-api-access-69hjq\") pod \"redhat-operators-xwscz\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.500031 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-utilities\") pod \"redhat-operators-xwscz\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.500057 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-catalog-content\") pod \"redhat-operators-xwscz\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.601322 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69hjq\" (UniqueName: \"kubernetes.io/projected/f120d8e7-a318-4238-ab2d-022b06d89ab7-kube-api-access-69hjq\") pod \"redhat-operators-xwscz\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.601437 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-utilities\") pod \"redhat-operators-xwscz\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.601462 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-catalog-content\") pod \"redhat-operators-xwscz\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.601900 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-catalog-content\") pod \"redhat-operators-xwscz\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.602470 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-utilities\") pod \"redhat-operators-xwscz\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.634754 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69hjq\" (UniqueName: \"kubernetes.io/projected/f120d8e7-a318-4238-ab2d-022b06d89ab7-kube-api-access-69hjq\") pod \"redhat-operators-xwscz\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.692859 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.954679 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwscz"] Dec 06 04:41:08 crc kubenswrapper[4718]: I1206 04:41:08.982031 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwscz" event={"ID":"f120d8e7-a318-4238-ab2d-022b06d89ab7","Type":"ContainerStarted","Data":"8f0237a13cde12a0f368031e91cb6922301ed1d257c87372f50e1bf133875b4e"} Dec 06 04:41:09 crc kubenswrapper[4718]: I1206 04:41:09.990911 4718 generic.go:334] "Generic (PLEG): container finished" podID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerID="ab7a545451a642804f352693782ea89d0f9f5c0a152cf3caf833be97c36f87dc" exitCode=0 Dec 06 04:41:09 crc kubenswrapper[4718]: I1206 04:41:09.991137 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwscz" event={"ID":"f120d8e7-a318-4238-ab2d-022b06d89ab7","Type":"ContainerDied","Data":"ab7a545451a642804f352693782ea89d0f9f5c0a152cf3caf833be97c36f87dc"} Dec 06 04:41:10 crc kubenswrapper[4718]: I1206 04:41:10.998734 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwscz" event={"ID":"f120d8e7-a318-4238-ab2d-022b06d89ab7","Type":"ContainerStarted","Data":"67f302f1537d25b8fca46aa7a6a1b844de76c494b7c8088bde612e0dac1e8404"} Dec 06 04:41:12 crc kubenswrapper[4718]: I1206 04:41:12.006105 4718 generic.go:334] "Generic (PLEG): container finished" podID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerID="67f302f1537d25b8fca46aa7a6a1b844de76c494b7c8088bde612e0dac1e8404" exitCode=0 Dec 06 04:41:12 crc kubenswrapper[4718]: I1206 04:41:12.006186 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwscz" event={"ID":"f120d8e7-a318-4238-ab2d-022b06d89ab7","Type":"ContainerDied","Data":"67f302f1537d25b8fca46aa7a6a1b844de76c494b7c8088bde612e0dac1e8404"} Dec 06 04:41:13 crc kubenswrapper[4718]: I1206 04:41:13.015313 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwscz" event={"ID":"f120d8e7-a318-4238-ab2d-022b06d89ab7","Type":"ContainerStarted","Data":"c8ba76ba56b5f0bbdb7789118e0727c8dccab3d0b77cde705e2e15a904c4370a"} Dec 06 04:41:13 crc kubenswrapper[4718]: I1206 04:41:13.043225 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xwscz" podStartSLOduration=2.646824445 podStartE2EDuration="5.043197539s" podCreationTimestamp="2025-12-06 04:41:08 +0000 UTC" firstStartedPulling="2025-12-06 04:41:09.992800311 +0000 UTC m=+2058.998505472" lastFinishedPulling="2025-12-06 04:41:12.389173405 +0000 UTC m=+2061.394878566" observedRunningTime="2025-12-06 04:41:13.028524857 +0000 UTC m=+2062.034230028" watchObservedRunningTime="2025-12-06 04:41:13.043197539 +0000 UTC m=+2062.048902710" Dec 06 04:41:15 crc kubenswrapper[4718]: I1206 04:41:15.956047 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5vd46/must-gather-vhz9z"] Dec 06 04:41:15 crc kubenswrapper[4718]: I1206 04:41:15.957843 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:41:15 crc kubenswrapper[4718]: I1206 04:41:15.961036 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5vd46"/"default-dockercfg-ncf9s" Dec 06 04:41:15 crc kubenswrapper[4718]: I1206 04:41:15.961355 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5vd46"/"openshift-service-ca.crt" Dec 06 04:41:15 crc kubenswrapper[4718]: I1206 04:41:15.961527 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5vd46"/"kube-root-ca.crt" Dec 06 04:41:15 crc kubenswrapper[4718]: I1206 04:41:15.973177 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5vd46/must-gather-vhz9z"] Dec 06 04:41:15 crc kubenswrapper[4718]: I1206 04:41:15.994091 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-must-gather-output\") pod \"must-gather-vhz9z\" (UID: \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\") " pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:41:15 crc kubenswrapper[4718]: I1206 04:41:15.994158 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d5jf\" (UniqueName: \"kubernetes.io/projected/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-kube-api-access-4d5jf\") pod \"must-gather-vhz9z\" (UID: \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\") " pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:41:16 crc kubenswrapper[4718]: I1206 04:41:16.095061 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-must-gather-output\") pod \"must-gather-vhz9z\" (UID: \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\") " pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:41:16 crc kubenswrapper[4718]: I1206 04:41:16.095147 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d5jf\" (UniqueName: \"kubernetes.io/projected/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-kube-api-access-4d5jf\") pod \"must-gather-vhz9z\" (UID: \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\") " pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:41:16 crc kubenswrapper[4718]: I1206 04:41:16.095697 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-must-gather-output\") pod \"must-gather-vhz9z\" (UID: \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\") " pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:41:16 crc kubenswrapper[4718]: I1206 04:41:16.113102 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d5jf\" (UniqueName: \"kubernetes.io/projected/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-kube-api-access-4d5jf\") pod \"must-gather-vhz9z\" (UID: \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\") " pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:41:16 crc kubenswrapper[4718]: I1206 04:41:16.274881 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:41:16 crc kubenswrapper[4718]: W1206 04:41:16.503917 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d86c587_ba7e_4c1a_b87c_e0a3d2de9fab.slice/crio-674eeaf57c78ffd52804225176f7b085745fe3ff2ea66154b5af45cd0d75bfb2 WatchSource:0}: Error finding container 674eeaf57c78ffd52804225176f7b085745fe3ff2ea66154b5af45cd0d75bfb2: Status 404 returned error can't find the container with id 674eeaf57c78ffd52804225176f7b085745fe3ff2ea66154b5af45cd0d75bfb2 Dec 06 04:41:16 crc kubenswrapper[4718]: I1206 04:41:16.504451 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5vd46/must-gather-vhz9z"] Dec 06 04:41:17 crc kubenswrapper[4718]: I1206 04:41:17.038094 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vd46/must-gather-vhz9z" event={"ID":"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab","Type":"ContainerStarted","Data":"674eeaf57c78ffd52804225176f7b085745fe3ff2ea66154b5af45cd0d75bfb2"} Dec 06 04:41:18 crc kubenswrapper[4718]: I1206 04:41:18.693902 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:18 crc kubenswrapper[4718]: I1206 04:41:18.694285 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:18 crc kubenswrapper[4718]: I1206 04:41:18.741654 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:19 crc kubenswrapper[4718]: I1206 04:41:19.103320 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:19 crc kubenswrapper[4718]: I1206 04:41:19.142511 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwscz"] Dec 06 04:41:21 crc kubenswrapper[4718]: I1206 04:41:21.076366 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xwscz" podUID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerName="registry-server" containerID="cri-o://c8ba76ba56b5f0bbdb7789118e0727c8dccab3d0b77cde705e2e15a904c4370a" gracePeriod=2 Dec 06 04:41:22 crc kubenswrapper[4718]: I1206 04:41:22.082766 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vd46/must-gather-vhz9z" event={"ID":"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab","Type":"ContainerStarted","Data":"6f18368bc59dc3f6b2f16d03f381bd231417914db48d6323c57f7258d3d9a5dc"} Dec 06 04:41:22 crc kubenswrapper[4718]: I1206 04:41:22.083039 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vd46/must-gather-vhz9z" event={"ID":"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab","Type":"ContainerStarted","Data":"97592df6292574768ce9ee66b00fe813b74502ddb1690548acb60dfe9a272dc4"} Dec 06 04:41:22 crc kubenswrapper[4718]: I1206 04:41:22.099067 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5vd46/must-gather-vhz9z" podStartSLOduration=2.8930684209999997 podStartE2EDuration="7.099052023s" podCreationTimestamp="2025-12-06 04:41:15 +0000 UTC" firstStartedPulling="2025-12-06 04:41:16.506575723 +0000 UTC m=+2065.512280884" lastFinishedPulling="2025-12-06 04:41:20.712559325 +0000 UTC m=+2069.718264486" observedRunningTime="2025-12-06 04:41:22.09698748 +0000 UTC m=+2071.102692651" watchObservedRunningTime="2025-12-06 04:41:22.099052023 +0000 UTC m=+2071.104757184" Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.103653 4718 generic.go:334] "Generic (PLEG): container finished" podID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerID="c8ba76ba56b5f0bbdb7789118e0727c8dccab3d0b77cde705e2e15a904c4370a" exitCode=0 Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.103706 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwscz" event={"ID":"f120d8e7-a318-4238-ab2d-022b06d89ab7","Type":"ContainerDied","Data":"c8ba76ba56b5f0bbdb7789118e0727c8dccab3d0b77cde705e2e15a904c4370a"} Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.221653 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.423465 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69hjq\" (UniqueName: \"kubernetes.io/projected/f120d8e7-a318-4238-ab2d-022b06d89ab7-kube-api-access-69hjq\") pod \"f120d8e7-a318-4238-ab2d-022b06d89ab7\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.423583 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-utilities\") pod \"f120d8e7-a318-4238-ab2d-022b06d89ab7\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.423623 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-catalog-content\") pod \"f120d8e7-a318-4238-ab2d-022b06d89ab7\" (UID: \"f120d8e7-a318-4238-ab2d-022b06d89ab7\") " Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.425505 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-utilities" (OuterVolumeSpecName: "utilities") pod "f120d8e7-a318-4238-ab2d-022b06d89ab7" (UID: "f120d8e7-a318-4238-ab2d-022b06d89ab7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.431418 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f120d8e7-a318-4238-ab2d-022b06d89ab7-kube-api-access-69hjq" (OuterVolumeSpecName: "kube-api-access-69hjq") pod "f120d8e7-a318-4238-ab2d-022b06d89ab7" (UID: "f120d8e7-a318-4238-ab2d-022b06d89ab7"). InnerVolumeSpecName "kube-api-access-69hjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.526031 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69hjq\" (UniqueName: \"kubernetes.io/projected/f120d8e7-a318-4238-ab2d-022b06d89ab7-kube-api-access-69hjq\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.526086 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.539358 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f120d8e7-a318-4238-ab2d-022b06d89ab7" (UID: "f120d8e7-a318-4238-ab2d-022b06d89ab7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:41:25 crc kubenswrapper[4718]: I1206 04:41:25.626973 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f120d8e7-a318-4238-ab2d-022b06d89ab7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:41:26 crc kubenswrapper[4718]: I1206 04:41:26.115463 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwscz" event={"ID":"f120d8e7-a318-4238-ab2d-022b06d89ab7","Type":"ContainerDied","Data":"8f0237a13cde12a0f368031e91cb6922301ed1d257c87372f50e1bf133875b4e"} Dec 06 04:41:26 crc kubenswrapper[4718]: I1206 04:41:26.115509 4718 scope.go:117] "RemoveContainer" containerID="c8ba76ba56b5f0bbdb7789118e0727c8dccab3d0b77cde705e2e15a904c4370a" Dec 06 04:41:26 crc kubenswrapper[4718]: I1206 04:41:26.115623 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwscz" Dec 06 04:41:26 crc kubenswrapper[4718]: I1206 04:41:26.139984 4718 scope.go:117] "RemoveContainer" containerID="67f302f1537d25b8fca46aa7a6a1b844de76c494b7c8088bde612e0dac1e8404" Dec 06 04:41:26 crc kubenswrapper[4718]: I1206 04:41:26.147308 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwscz"] Dec 06 04:41:26 crc kubenswrapper[4718]: I1206 04:41:26.162541 4718 scope.go:117] "RemoveContainer" containerID="ab7a545451a642804f352693782ea89d0f9f5c0a152cf3caf833be97c36f87dc" Dec 06 04:41:26 crc kubenswrapper[4718]: I1206 04:41:26.167619 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xwscz"] Dec 06 04:41:27 crc kubenswrapper[4718]: I1206 04:41:27.335800 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f120d8e7-a318-4238-ab2d-022b06d89ab7" path="/var/lib/kubelet/pods/f120d8e7-a318-4238-ab2d-022b06d89ab7/volumes" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.345328 4718 scope.go:117] "RemoveContainer" containerID="5de89b4026a57af3c1e76a6b8ef008b05b02e357800f04d707cf96ee442ce5e4" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.369686 4718 scope.go:117] "RemoveContainer" containerID="8a56447d0b735769e654f4b328cc14bd65ec17aa99f7bdb216d0fb0d427661fd" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.395023 4718 scope.go:117] "RemoveContainer" containerID="b9bb347c8c4f66de0bbffb988d1988ce2defc3534b84539638e8a0d80a129ceb" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.425620 4718 scope.go:117] "RemoveContainer" containerID="715de72658276c0e32f61d47b2bec2d5a432ab5fd4fd2816bfe6e6b17d4ba802" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.441375 4718 scope.go:117] "RemoveContainer" containerID="2d1388cbab7597dc1d9bd2d00889495b9b31924c1932ca8836a9d9ccf35076f8" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.464033 4718 scope.go:117] "RemoveContainer" containerID="3f8b90b980ff5b05fdca17e34d36049ee00b9960b8169dfe6a5c32b0e52bc034" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.481966 4718 scope.go:117] "RemoveContainer" containerID="6bfed44d0979e803bc0093f058cc12be1bdcf12fff577d2c92d5fbe36eadd256" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.495204 4718 scope.go:117] "RemoveContainer" containerID="bd262fbe949bba5ca68d8d315477d0ac3f1fd75b84cb448db0cd1a01602fc00a" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.511645 4718 scope.go:117] "RemoveContainer" containerID="21cf09f334cc01a5c689c212a101b15b4d768daa5648ebbe24dc25314d337ab7" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.525948 4718 scope.go:117] "RemoveContainer" containerID="61b2860f6878a3eb87308c49873c973e01b93112967cfd54ab36e38fdcc98475" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.549193 4718 scope.go:117] "RemoveContainer" containerID="2cdfdb17b2603828d33a7279cdb768c507d3dadec658af3c6a3cf287a1d4df22" Dec 06 04:41:53 crc kubenswrapper[4718]: I1206 04:41:53.566988 4718 scope.go:117] "RemoveContainer" containerID="82e89a00ff8b153a78b8b13c5f8c6095db280876aa36864f4caa0d42af26654a" Dec 06 04:42:03 crc kubenswrapper[4718]: I1206 04:42:03.692054 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-g2qsr_96226f3a-94e6-4895-a9fc-662f07f8cdf8/control-plane-machine-set-operator/0.log" Dec 06 04:42:03 crc kubenswrapper[4718]: I1206 04:42:03.791616 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6jmn4_5d02801a-a515-4f1f-99a5-90863caec537/kube-rbac-proxy/0.log" Dec 06 04:42:03 crc kubenswrapper[4718]: I1206 04:42:03.845989 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6jmn4_5d02801a-a515-4f1f-99a5-90863caec537/machine-api-operator/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.289134 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-r6chz_b00ca948-0cd8-43bc-8b38-43f445044e51/controller/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.300649 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-r6chz_b00ca948-0cd8-43bc-8b38-43f445044e51/kube-rbac-proxy/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.457637 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-frr-files/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.633707 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-reloader/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.656930 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-reloader/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.657831 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-frr-files/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.668599 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-metrics/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.827854 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-frr-files/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.838761 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-metrics/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.868028 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-metrics/0.log" Dec 06 04:42:19 crc kubenswrapper[4718]: I1206 04:42:19.881821 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-reloader/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.136493 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-reloader/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.155001 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-frr-files/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.158881 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-metrics/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.163183 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/controller/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.362654 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/kube-rbac-proxy/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.385291 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/frr-metrics/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.405756 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/kube-rbac-proxy-frr/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.556466 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-7swws_0b00eded-0715-4a41-89e9-3f758847a16a/frr-k8s-webhook-server/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.564793 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/reloader/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.686838 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/frr/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.776393 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-57d558fdb8-rqqm5_939e915d-3ad4-4a15-bb70-01106cd5cb96/manager/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.853429 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5c49888cd9-w2wwt_e5e44d0d-be18-4944-ae68-8758042af15e/webhook-server/0.log" Dec 06 04:42:20 crc kubenswrapper[4718]: I1206 04:42:20.940554 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9wqkt_1bc1c47e-1be5-4a69-96cf-fa5c59004f8b/kube-rbac-proxy/0.log" Dec 06 04:42:21 crc kubenswrapper[4718]: I1206 04:42:21.025522 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9wqkt_1bc1c47e-1be5-4a69-96cf-fa5c59004f8b/speaker/0.log" Dec 06 04:42:27 crc kubenswrapper[4718]: I1206 04:42:27.877604 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:42:27 crc kubenswrapper[4718]: I1206 04:42:27.878174 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:42:44 crc kubenswrapper[4718]: I1206 04:42:44.922704 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/util/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.035354 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/util/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.081938 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/pull/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.101423 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/pull/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.289677 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/util/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.299700 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/pull/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.326391 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/extract/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.449887 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-utilities/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.575690 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-content/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.599742 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-utilities/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.610789 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-content/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.769401 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-utilities/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.780652 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-content/0.log" Dec 06 04:42:45 crc kubenswrapper[4718]: I1206 04:42:45.970342 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-utilities/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.112486 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-content/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.135792 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/registry-server/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.151613 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-utilities/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.165675 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-content/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.333406 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-utilities/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.355174 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-content/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.592027 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-42x7c_49e186cc-90d7-4d3e-973b-1727f13a1c6d/marketplace-operator/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.618713 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-utilities/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.786445 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/registry-server/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.803261 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-utilities/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.833705 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-content/0.log" Dec 06 04:42:46 crc kubenswrapper[4718]: I1206 04:42:46.885150 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-content/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.060632 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-utilities/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.116881 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-content/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.187022 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/registry-server/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.240423 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-utilities/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.417053 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-utilities/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.420874 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-content/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.446401 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-content/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.613345 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-utilities/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.633267 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-content/0.log" Dec 06 04:42:47 crc kubenswrapper[4718]: I1206 04:42:47.951594 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/registry-server/0.log" Dec 06 04:42:57 crc kubenswrapper[4718]: I1206 04:42:57.877268 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:42:57 crc kubenswrapper[4718]: I1206 04:42:57.877898 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:43:27 crc kubenswrapper[4718]: I1206 04:43:27.877781 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:43:27 crc kubenswrapper[4718]: I1206 04:43:27.878398 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:43:27 crc kubenswrapper[4718]: I1206 04:43:27.878451 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:43:27 crc kubenswrapper[4718]: I1206 04:43:27.878981 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae94ee04e0f0b6f53a2fb12aac1d020fc0dc1260d9f11235b84486c47305e1c7"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:43:27 crc kubenswrapper[4718]: I1206 04:43:27.879035 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://ae94ee04e0f0b6f53a2fb12aac1d020fc0dc1260d9f11235b84486c47305e1c7" gracePeriod=600 Dec 06 04:43:28 crc kubenswrapper[4718]: I1206 04:43:28.850579 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="ae94ee04e0f0b6f53a2fb12aac1d020fc0dc1260d9f11235b84486c47305e1c7" exitCode=0 Dec 06 04:43:28 crc kubenswrapper[4718]: I1206 04:43:28.850673 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"ae94ee04e0f0b6f53a2fb12aac1d020fc0dc1260d9f11235b84486c47305e1c7"} Dec 06 04:43:28 crc kubenswrapper[4718]: I1206 04:43:28.851512 4718 scope.go:117] "RemoveContainer" containerID="280c459e8c9d2173a56230153bbd9dd327a00e007275303c6395fd45f01d1937" Dec 06 04:43:29 crc kubenswrapper[4718]: I1206 04:43:29.861393 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037"} Dec 06 04:43:52 crc kubenswrapper[4718]: I1206 04:43:52.008749 4718 generic.go:334] "Generic (PLEG): container finished" podID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" containerID="6f18368bc59dc3f6b2f16d03f381bd231417914db48d6323c57f7258d3d9a5dc" exitCode=0 Dec 06 04:43:52 crc kubenswrapper[4718]: I1206 04:43:52.008929 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vd46/must-gather-vhz9z" event={"ID":"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab","Type":"ContainerDied","Data":"6f18368bc59dc3f6b2f16d03f381bd231417914db48d6323c57f7258d3d9a5dc"} Dec 06 04:43:52 crc kubenswrapper[4718]: I1206 04:43:52.010530 4718 scope.go:117] "RemoveContainer" containerID="6f18368bc59dc3f6b2f16d03f381bd231417914db48d6323c57f7258d3d9a5dc" Dec 06 04:43:52 crc kubenswrapper[4718]: I1206 04:43:52.506828 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5vd46_must-gather-vhz9z_0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab/gather/0.log" Dec 06 04:43:59 crc kubenswrapper[4718]: I1206 04:43:59.693842 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5vd46/must-gather-vhz9z"] Dec 06 04:43:59 crc kubenswrapper[4718]: I1206 04:43:59.694513 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-5vd46/must-gather-vhz9z" podUID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" containerName="copy" containerID="cri-o://97592df6292574768ce9ee66b00fe813b74502ddb1690548acb60dfe9a272dc4" gracePeriod=2 Dec 06 04:43:59 crc kubenswrapper[4718]: I1206 04:43:59.697410 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5vd46/must-gather-vhz9z"] Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.059506 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5vd46_must-gather-vhz9z_0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab/copy/0.log" Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.060154 4718 generic.go:334] "Generic (PLEG): container finished" podID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" containerID="97592df6292574768ce9ee66b00fe813b74502ddb1690548acb60dfe9a272dc4" exitCode=143 Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.060204 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="674eeaf57c78ffd52804225176f7b085745fe3ff2ea66154b5af45cd0d75bfb2" Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.073675 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5vd46_must-gather-vhz9z_0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab/copy/0.log" Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.074083 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.235926 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d5jf\" (UniqueName: \"kubernetes.io/projected/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-kube-api-access-4d5jf\") pod \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\" (UID: \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\") " Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.236082 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-must-gather-output\") pod \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\" (UID: \"0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab\") " Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.241989 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-kube-api-access-4d5jf" (OuterVolumeSpecName: "kube-api-access-4d5jf") pod "0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" (UID: "0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab"). InnerVolumeSpecName "kube-api-access-4d5jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.328125 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" (UID: "0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.337936 4718 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 04:44:00 crc kubenswrapper[4718]: I1206 04:44:00.337977 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d5jf\" (UniqueName: \"kubernetes.io/projected/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab-kube-api-access-4d5jf\") on node \"crc\" DevicePath \"\"" Dec 06 04:44:01 crc kubenswrapper[4718]: I1206 04:44:01.064658 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vd46/must-gather-vhz9z" Dec 06 04:44:01 crc kubenswrapper[4718]: I1206 04:44:01.338350 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" path="/var/lib/kubelet/pods/0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab/volumes" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.028703 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cprcd"] Dec 06 04:44:14 crc kubenswrapper[4718]: E1206 04:44:14.030010 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerName="extract-utilities" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.030066 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerName="extract-utilities" Dec 06 04:44:14 crc kubenswrapper[4718]: E1206 04:44:14.030097 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerName="extract-content" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.030117 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerName="extract-content" Dec 06 04:44:14 crc kubenswrapper[4718]: E1206 04:44:14.030150 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" containerName="gather" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.030170 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" containerName="gather" Dec 06 04:44:14 crc kubenswrapper[4718]: E1206 04:44:14.030198 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" containerName="copy" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.030218 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" containerName="copy" Dec 06 04:44:14 crc kubenswrapper[4718]: E1206 04:44:14.030294 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerName="registry-server" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.030314 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerName="registry-server" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.030647 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" containerName="gather" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.030676 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d86c587-ba7e-4c1a-b87c-e0a3d2de9fab" containerName="copy" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.030726 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="f120d8e7-a318-4238-ab2d-022b06d89ab7" containerName="registry-server" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.032527 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cprcd"] Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.032677 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.040877 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-catalog-content\") pod \"community-operators-cprcd\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.040966 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-utilities\") pod \"community-operators-cprcd\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.041115 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjxx4\" (UniqueName: \"kubernetes.io/projected/ce09672f-e117-41bf-a98c-f08a28b9a711-kube-api-access-sjxx4\") pod \"community-operators-cprcd\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.141668 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-utilities\") pod \"community-operators-cprcd\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.141793 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjxx4\" (UniqueName: \"kubernetes.io/projected/ce09672f-e117-41bf-a98c-f08a28b9a711-kube-api-access-sjxx4\") pod \"community-operators-cprcd\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.141836 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-catalog-content\") pod \"community-operators-cprcd\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.142127 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-utilities\") pod \"community-operators-cprcd\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.142438 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-catalog-content\") pod \"community-operators-cprcd\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.172618 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjxx4\" (UniqueName: \"kubernetes.io/projected/ce09672f-e117-41bf-a98c-f08a28b9a711-kube-api-access-sjxx4\") pod \"community-operators-cprcd\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.352101 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:14 crc kubenswrapper[4718]: I1206 04:44:14.621370 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cprcd"] Dec 06 04:44:14 crc kubenswrapper[4718]: W1206 04:44:14.628913 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce09672f_e117_41bf_a98c_f08a28b9a711.slice/crio-45d5ed7ef5cbf46e31fe89d684383fd472f4dc9660cbcfdf3e284525bd5301ba WatchSource:0}: Error finding container 45d5ed7ef5cbf46e31fe89d684383fd472f4dc9660cbcfdf3e284525bd5301ba: Status 404 returned error can't find the container with id 45d5ed7ef5cbf46e31fe89d684383fd472f4dc9660cbcfdf3e284525bd5301ba Dec 06 04:44:15 crc kubenswrapper[4718]: I1206 04:44:15.149193 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cprcd" event={"ID":"ce09672f-e117-41bf-a98c-f08a28b9a711","Type":"ContainerStarted","Data":"45d5ed7ef5cbf46e31fe89d684383fd472f4dc9660cbcfdf3e284525bd5301ba"} Dec 06 04:44:16 crc kubenswrapper[4718]: I1206 04:44:16.157403 4718 generic.go:334] "Generic (PLEG): container finished" podID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerID="864a0bae93c75d8745d6d25c800dfa3a328675db2c388c8e43806170dff9b0dd" exitCode=0 Dec 06 04:44:16 crc kubenswrapper[4718]: I1206 04:44:16.157479 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cprcd" event={"ID":"ce09672f-e117-41bf-a98c-f08a28b9a711","Type":"ContainerDied","Data":"864a0bae93c75d8745d6d25c800dfa3a328675db2c388c8e43806170dff9b0dd"} Dec 06 04:44:18 crc kubenswrapper[4718]: I1206 04:44:18.182273 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cprcd" event={"ID":"ce09672f-e117-41bf-a98c-f08a28b9a711","Type":"ContainerStarted","Data":"720012f62ab9f7c425b308d2c33dd08dcd6d1a54afc45f00e224aae85f255363"} Dec 06 04:44:19 crc kubenswrapper[4718]: I1206 04:44:19.192052 4718 generic.go:334] "Generic (PLEG): container finished" podID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerID="720012f62ab9f7c425b308d2c33dd08dcd6d1a54afc45f00e224aae85f255363" exitCode=0 Dec 06 04:44:19 crc kubenswrapper[4718]: I1206 04:44:19.192123 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cprcd" event={"ID":"ce09672f-e117-41bf-a98c-f08a28b9a711","Type":"ContainerDied","Data":"720012f62ab9f7c425b308d2c33dd08dcd6d1a54afc45f00e224aae85f255363"} Dec 06 04:44:21 crc kubenswrapper[4718]: I1206 04:44:21.209207 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cprcd" event={"ID":"ce09672f-e117-41bf-a98c-f08a28b9a711","Type":"ContainerStarted","Data":"1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af"} Dec 06 04:44:21 crc kubenswrapper[4718]: I1206 04:44:21.236882 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cprcd" podStartSLOduration=3.037251099 podStartE2EDuration="7.236863133s" podCreationTimestamp="2025-12-06 04:44:14 +0000 UTC" firstStartedPulling="2025-12-06 04:44:16.159484565 +0000 UTC m=+2245.165189746" lastFinishedPulling="2025-12-06 04:44:20.359096619 +0000 UTC m=+2249.364801780" observedRunningTime="2025-12-06 04:44:21.231793684 +0000 UTC m=+2250.237498845" watchObservedRunningTime="2025-12-06 04:44:21.236863133 +0000 UTC m=+2250.242568314" Dec 06 04:44:24 crc kubenswrapper[4718]: I1206 04:44:24.352485 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:24 crc kubenswrapper[4718]: I1206 04:44:24.352848 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:24 crc kubenswrapper[4718]: I1206 04:44:24.402006 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:25 crc kubenswrapper[4718]: I1206 04:44:25.310517 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:25 crc kubenswrapper[4718]: I1206 04:44:25.372273 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cprcd"] Dec 06 04:44:27 crc kubenswrapper[4718]: I1206 04:44:27.246672 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cprcd" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerName="registry-server" containerID="cri-o://1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af" gracePeriod=2 Dec 06 04:44:34 crc kubenswrapper[4718]: E1206 04:44:34.353132 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af is running failed: container process not found" containerID="1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 04:44:34 crc kubenswrapper[4718]: E1206 04:44:34.353994 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af is running failed: container process not found" containerID="1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 04:44:34 crc kubenswrapper[4718]: E1206 04:44:34.354192 4718 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af is running failed: container process not found" containerID="1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 04:44:34 crc kubenswrapper[4718]: E1206 04:44:34.354261 4718 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-cprcd" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerName="registry-server" Dec 06 04:44:36 crc kubenswrapper[4718]: I1206 04:44:36.864686 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cprcd_ce09672f-e117-41bf-a98c-f08a28b9a711/registry-server/0.log" Dec 06 04:44:36 crc kubenswrapper[4718]: I1206 04:44:36.867758 4718 generic.go:334] "Generic (PLEG): container finished" podID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerID="1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af" exitCode=-1 Dec 06 04:44:36 crc kubenswrapper[4718]: I1206 04:44:36.867837 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cprcd" event={"ID":"ce09672f-e117-41bf-a98c-f08a28b9a711","Type":"ContainerDied","Data":"1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af"} Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.103596 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.251173 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-utilities\") pod \"ce09672f-e117-41bf-a98c-f08a28b9a711\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.251290 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-catalog-content\") pod \"ce09672f-e117-41bf-a98c-f08a28b9a711\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.252316 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-utilities" (OuterVolumeSpecName: "utilities") pod "ce09672f-e117-41bf-a98c-f08a28b9a711" (UID: "ce09672f-e117-41bf-a98c-f08a28b9a711"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.252482 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjxx4\" (UniqueName: \"kubernetes.io/projected/ce09672f-e117-41bf-a98c-f08a28b9a711-kube-api-access-sjxx4\") pod \"ce09672f-e117-41bf-a98c-f08a28b9a711\" (UID: \"ce09672f-e117-41bf-a98c-f08a28b9a711\") " Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.252736 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.260973 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce09672f-e117-41bf-a98c-f08a28b9a711-kube-api-access-sjxx4" (OuterVolumeSpecName: "kube-api-access-sjxx4") pod "ce09672f-e117-41bf-a98c-f08a28b9a711" (UID: "ce09672f-e117-41bf-a98c-f08a28b9a711"). InnerVolumeSpecName "kube-api-access-sjxx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.312813 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce09672f-e117-41bf-a98c-f08a28b9a711" (UID: "ce09672f-e117-41bf-a98c-f08a28b9a711"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.353743 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce09672f-e117-41bf-a98c-f08a28b9a711-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.353977 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjxx4\" (UniqueName: \"kubernetes.io/projected/ce09672f-e117-41bf-a98c-f08a28b9a711-kube-api-access-sjxx4\") on node \"crc\" DevicePath \"\"" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.881519 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cprcd" event={"ID":"ce09672f-e117-41bf-a98c-f08a28b9a711","Type":"ContainerDied","Data":"45d5ed7ef5cbf46e31fe89d684383fd472f4dc9660cbcfdf3e284525bd5301ba"} Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.881592 4718 scope.go:117] "RemoveContainer" containerID="1f02b125b3714e865bf12b32bd6d4ef19fddf6158eb71dfbe6eeeb8479ca17af" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.881702 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cprcd" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.902084 4718 scope.go:117] "RemoveContainer" containerID="720012f62ab9f7c425b308d2c33dd08dcd6d1a54afc45f00e224aae85f255363" Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.914400 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cprcd"] Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.920713 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cprcd"] Dec 06 04:44:37 crc kubenswrapper[4718]: I1206 04:44:37.933491 4718 scope.go:117] "RemoveContainer" containerID="864a0bae93c75d8745d6d25c800dfa3a328675db2c388c8e43806170dff9b0dd" Dec 06 04:44:39 crc kubenswrapper[4718]: I1206 04:44:39.334765 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" path="/var/lib/kubelet/pods/ce09672f-e117-41bf-a98c-f08a28b9a711/volumes" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.139570 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq"] Dec 06 04:45:00 crc kubenswrapper[4718]: E1206 04:45:00.140635 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerName="registry-server" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.140665 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerName="registry-server" Dec 06 04:45:00 crc kubenswrapper[4718]: E1206 04:45:00.140695 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerName="extract-utilities" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.140709 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerName="extract-utilities" Dec 06 04:45:00 crc kubenswrapper[4718]: E1206 04:45:00.140742 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerName="extract-content" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.140757 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerName="extract-content" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.140937 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce09672f-e117-41bf-a98c-f08a28b9a711" containerName="registry-server" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.141583 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.147146 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.147263 4718 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.153856 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq"] Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.260561 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-config-volume\") pod \"collect-profiles-29416605-8mqnq\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.260639 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk7xh\" (UniqueName: \"kubernetes.io/projected/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-kube-api-access-hk7xh\") pod \"collect-profiles-29416605-8mqnq\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.260671 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-secret-volume\") pod \"collect-profiles-29416605-8mqnq\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.362210 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-config-volume\") pod \"collect-profiles-29416605-8mqnq\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.362376 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk7xh\" (UniqueName: \"kubernetes.io/projected/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-kube-api-access-hk7xh\") pod \"collect-profiles-29416605-8mqnq\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.362407 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-secret-volume\") pod \"collect-profiles-29416605-8mqnq\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.364029 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-config-volume\") pod \"collect-profiles-29416605-8mqnq\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.368323 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-secret-volume\") pod \"collect-profiles-29416605-8mqnq\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.383258 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk7xh\" (UniqueName: \"kubernetes.io/projected/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-kube-api-access-hk7xh\") pod \"collect-profiles-29416605-8mqnq\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.458625 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:00 crc kubenswrapper[4718]: I1206 04:45:00.906128 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq"] Dec 06 04:45:01 crc kubenswrapper[4718]: I1206 04:45:01.032210 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" event={"ID":"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1","Type":"ContainerStarted","Data":"9ace630b361c5660eaae9f0c63ec4afe22f6b3537e9c95631554f21ce51ca834"} Dec 06 04:45:02 crc kubenswrapper[4718]: I1206 04:45:02.038500 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" event={"ID":"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1","Type":"ContainerStarted","Data":"02b6c17cd85c2f21047d3a8c36aad921f1ab2be88293e4e2f07d5bbaf4920b96"} Dec 06 04:45:02 crc kubenswrapper[4718]: I1206 04:45:02.054470 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" podStartSLOduration=2.054451295 podStartE2EDuration="2.054451295s" podCreationTimestamp="2025-12-06 04:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:45:02.054385864 +0000 UTC m=+2291.060091025" watchObservedRunningTime="2025-12-06 04:45:02.054451295 +0000 UTC m=+2291.060156456" Dec 06 04:45:03 crc kubenswrapper[4718]: I1206 04:45:03.048090 4718 generic.go:334] "Generic (PLEG): container finished" podID="5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1" containerID="02b6c17cd85c2f21047d3a8c36aad921f1ab2be88293e4e2f07d5bbaf4920b96" exitCode=0 Dec 06 04:45:03 crc kubenswrapper[4718]: I1206 04:45:03.048152 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" event={"ID":"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1","Type":"ContainerDied","Data":"02b6c17cd85c2f21047d3a8c36aad921f1ab2be88293e4e2f07d5bbaf4920b96"} Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.304760 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.417404 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-secret-volume\") pod \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.417553 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk7xh\" (UniqueName: \"kubernetes.io/projected/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-kube-api-access-hk7xh\") pod \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.417589 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-config-volume\") pod \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\" (UID: \"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1\") " Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.418089 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-config-volume" (OuterVolumeSpecName: "config-volume") pod "5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1" (UID: "5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.418994 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm"] Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.423409 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-kube-api-access-hk7xh" (OuterVolumeSpecName: "kube-api-access-hk7xh") pod "5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1" (UID: "5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1"). InnerVolumeSpecName "kube-api-access-hk7xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.423411 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1" (UID: "5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.426849 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416560-c95jm"] Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.518545 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk7xh\" (UniqueName: \"kubernetes.io/projected/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-kube-api-access-hk7xh\") on node \"crc\" DevicePath \"\"" Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.518918 4718 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:45:04 crc kubenswrapper[4718]: I1206 04:45:04.519014 4718 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 04:45:05 crc kubenswrapper[4718]: I1206 04:45:05.063717 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" event={"ID":"5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1","Type":"ContainerDied","Data":"9ace630b361c5660eaae9f0c63ec4afe22f6b3537e9c95631554f21ce51ca834"} Dec 06 04:45:05 crc kubenswrapper[4718]: I1206 04:45:05.063774 4718 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ace630b361c5660eaae9f0c63ec4afe22f6b3537e9c95631554f21ce51ca834" Dec 06 04:45:05 crc kubenswrapper[4718]: I1206 04:45:05.063773 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416605-8mqnq" Dec 06 04:45:05 crc kubenswrapper[4718]: I1206 04:45:05.340086 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f68d477b-1cc9-4ab9-8fa1-16fc30c9a845" path="/var/lib/kubelet/pods/f68d477b-1cc9-4ab9-8fa1-16fc30c9a845/volumes" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.281160 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnw7"] Dec 06 04:45:16 crc kubenswrapper[4718]: E1206 04:45:16.282132 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1" containerName="collect-profiles" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.282190 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1" containerName="collect-profiles" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.282438 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d5bdee2-13f8-4d32-9dd2-5fa2c60c7fc1" containerName="collect-profiles" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.283517 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.287974 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnw7"] Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.395034 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-utilities\") pod \"redhat-marketplace-dsnw7\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.395139 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-catalog-content\") pod \"redhat-marketplace-dsnw7\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.395196 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5ggn\" (UniqueName: \"kubernetes.io/projected/874a583c-dd2b-46f4-8777-1a66b2eb413a-kube-api-access-t5ggn\") pod \"redhat-marketplace-dsnw7\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.496878 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-catalog-content\") pod \"redhat-marketplace-dsnw7\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.496956 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5ggn\" (UniqueName: \"kubernetes.io/projected/874a583c-dd2b-46f4-8777-1a66b2eb413a-kube-api-access-t5ggn\") pod \"redhat-marketplace-dsnw7\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.497012 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-utilities\") pod \"redhat-marketplace-dsnw7\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.497619 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-catalog-content\") pod \"redhat-marketplace-dsnw7\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.497660 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-utilities\") pod \"redhat-marketplace-dsnw7\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.516017 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5ggn\" (UniqueName: \"kubernetes.io/projected/874a583c-dd2b-46f4-8777-1a66b2eb413a-kube-api-access-t5ggn\") pod \"redhat-marketplace-dsnw7\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.604883 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:16 crc kubenswrapper[4718]: I1206 04:45:16.802609 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnw7"] Dec 06 04:45:17 crc kubenswrapper[4718]: I1206 04:45:17.193037 4718 generic.go:334] "Generic (PLEG): container finished" podID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerID="8aa98602a1c0f2d6c8e7e6b77a49d24de5354acc4cc638c3fd4652a00c2401e3" exitCode=0 Dec 06 04:45:17 crc kubenswrapper[4718]: I1206 04:45:17.193104 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnw7" event={"ID":"874a583c-dd2b-46f4-8777-1a66b2eb413a","Type":"ContainerDied","Data":"8aa98602a1c0f2d6c8e7e6b77a49d24de5354acc4cc638c3fd4652a00c2401e3"} Dec 06 04:45:17 crc kubenswrapper[4718]: I1206 04:45:17.193165 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnw7" event={"ID":"874a583c-dd2b-46f4-8777-1a66b2eb413a","Type":"ContainerStarted","Data":"7ba788de7a0baa77e90fe554f068fe0f1a98379cb59236403b51a6611198a461"} Dec 06 04:45:20 crc kubenswrapper[4718]: I1206 04:45:20.215885 4718 generic.go:334] "Generic (PLEG): container finished" podID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerID="73bb5d4fcabc7f7e64c02caacddd75b06db0afac40e27a7b787644b7f1ecfec2" exitCode=0 Dec 06 04:45:20 crc kubenswrapper[4718]: I1206 04:45:20.215999 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnw7" event={"ID":"874a583c-dd2b-46f4-8777-1a66b2eb413a","Type":"ContainerDied","Data":"73bb5d4fcabc7f7e64c02caacddd75b06db0afac40e27a7b787644b7f1ecfec2"} Dec 06 04:45:22 crc kubenswrapper[4718]: I1206 04:45:22.233817 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnw7" event={"ID":"874a583c-dd2b-46f4-8777-1a66b2eb413a","Type":"ContainerStarted","Data":"bdab6f9e378ed9bb900eb096253d9d242c4e4bf95872bcdb9ea67ef6fe696ffc"} Dec 06 04:45:22 crc kubenswrapper[4718]: I1206 04:45:22.264360 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dsnw7" podStartSLOduration=2.375135714 podStartE2EDuration="6.264344041s" podCreationTimestamp="2025-12-06 04:45:16 +0000 UTC" firstStartedPulling="2025-12-06 04:45:17.194896074 +0000 UTC m=+2306.200601255" lastFinishedPulling="2025-12-06 04:45:21.084104411 +0000 UTC m=+2310.089809582" observedRunningTime="2025-12-06 04:45:22.257961059 +0000 UTC m=+2311.263666260" watchObservedRunningTime="2025-12-06 04:45:22.264344041 +0000 UTC m=+2311.270049202" Dec 06 04:45:26 crc kubenswrapper[4718]: I1206 04:45:26.605789 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:26 crc kubenswrapper[4718]: I1206 04:45:26.606449 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:26 crc kubenswrapper[4718]: I1206 04:45:26.664005 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:27 crc kubenswrapper[4718]: I1206 04:45:27.324815 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:27 crc kubenswrapper[4718]: I1206 04:45:27.379959 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnw7"] Dec 06 04:45:29 crc kubenswrapper[4718]: I1206 04:45:29.288183 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dsnw7" podUID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerName="registry-server" containerID="cri-o://bdab6f9e378ed9bb900eb096253d9d242c4e4bf95872bcdb9ea67ef6fe696ffc" gracePeriod=2 Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.308765 4718 generic.go:334] "Generic (PLEG): container finished" podID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerID="bdab6f9e378ed9bb900eb096253d9d242c4e4bf95872bcdb9ea67ef6fe696ffc" exitCode=0 Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.308866 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnw7" event={"ID":"874a583c-dd2b-46f4-8777-1a66b2eb413a","Type":"ContainerDied","Data":"bdab6f9e378ed9bb900eb096253d9d242c4e4bf95872bcdb9ea67ef6fe696ffc"} Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.537455 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.711287 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-catalog-content\") pod \"874a583c-dd2b-46f4-8777-1a66b2eb413a\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.711476 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5ggn\" (UniqueName: \"kubernetes.io/projected/874a583c-dd2b-46f4-8777-1a66b2eb413a-kube-api-access-t5ggn\") pod \"874a583c-dd2b-46f4-8777-1a66b2eb413a\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.711527 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-utilities\") pod \"874a583c-dd2b-46f4-8777-1a66b2eb413a\" (UID: \"874a583c-dd2b-46f4-8777-1a66b2eb413a\") " Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.713138 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-utilities" (OuterVolumeSpecName: "utilities") pod "874a583c-dd2b-46f4-8777-1a66b2eb413a" (UID: "874a583c-dd2b-46f4-8777-1a66b2eb413a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.721501 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/874a583c-dd2b-46f4-8777-1a66b2eb413a-kube-api-access-t5ggn" (OuterVolumeSpecName: "kube-api-access-t5ggn") pod "874a583c-dd2b-46f4-8777-1a66b2eb413a" (UID: "874a583c-dd2b-46f4-8777-1a66b2eb413a"). InnerVolumeSpecName "kube-api-access-t5ggn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.752878 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "874a583c-dd2b-46f4-8777-1a66b2eb413a" (UID: "874a583c-dd2b-46f4-8777-1a66b2eb413a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.814186 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5ggn\" (UniqueName: \"kubernetes.io/projected/874a583c-dd2b-46f4-8777-1a66b2eb413a-kube-api-access-t5ggn\") on node \"crc\" DevicePath \"\"" Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.814259 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:45:31 crc kubenswrapper[4718]: I1206 04:45:31.814274 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874a583c-dd2b-46f4-8777-1a66b2eb413a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:45:32 crc kubenswrapper[4718]: I1206 04:45:32.316873 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnw7" event={"ID":"874a583c-dd2b-46f4-8777-1a66b2eb413a","Type":"ContainerDied","Data":"7ba788de7a0baa77e90fe554f068fe0f1a98379cb59236403b51a6611198a461"} Dec 06 04:45:32 crc kubenswrapper[4718]: I1206 04:45:32.316929 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsnw7" Dec 06 04:45:32 crc kubenswrapper[4718]: I1206 04:45:32.317290 4718 scope.go:117] "RemoveContainer" containerID="bdab6f9e378ed9bb900eb096253d9d242c4e4bf95872bcdb9ea67ef6fe696ffc" Dec 06 04:45:32 crc kubenswrapper[4718]: I1206 04:45:32.336812 4718 scope.go:117] "RemoveContainer" containerID="73bb5d4fcabc7f7e64c02caacddd75b06db0afac40e27a7b787644b7f1ecfec2" Dec 06 04:45:32 crc kubenswrapper[4718]: I1206 04:45:32.352792 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnw7"] Dec 06 04:45:32 crc kubenswrapper[4718]: I1206 04:45:32.359725 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnw7"] Dec 06 04:45:32 crc kubenswrapper[4718]: I1206 04:45:32.372681 4718 scope.go:117] "RemoveContainer" containerID="8aa98602a1c0f2d6c8e7e6b77a49d24de5354acc4cc638c3fd4652a00c2401e3" Dec 06 04:45:33 crc kubenswrapper[4718]: I1206 04:45:33.338710 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="874a583c-dd2b-46f4-8777-1a66b2eb413a" path="/var/lib/kubelet/pods/874a583c-dd2b-46f4-8777-1a66b2eb413a/volumes" Dec 06 04:45:53 crc kubenswrapper[4718]: I1206 04:45:53.829893 4718 scope.go:117] "RemoveContainer" containerID="befec181c9ee2485645873c1b4022b35d584b42fbd0849eb58d281af10627b81" Dec 06 04:45:57 crc kubenswrapper[4718]: I1206 04:45:57.876769 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:45:57 crc kubenswrapper[4718]: I1206 04:45:57.877115 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.966270 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hgpvg/must-gather-4hdfc"] Dec 06 04:46:26 crc kubenswrapper[4718]: E1206 04:46:26.968346 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerName="extract-content" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.968485 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerName="extract-content" Dec 06 04:46:26 crc kubenswrapper[4718]: E1206 04:46:26.968616 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerName="registry-server" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.968723 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerName="registry-server" Dec 06 04:46:26 crc kubenswrapper[4718]: E1206 04:46:26.968850 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerName="extract-utilities" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.968989 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerName="extract-utilities" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.969361 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="874a583c-dd2b-46f4-8777-1a66b2eb413a" containerName="registry-server" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.970588 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.974690 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hgpvg"/"openshift-service-ca.crt" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.975021 4718 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hgpvg"/"kube-root-ca.crt" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.975057 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ef1b93f-22cf-446a-8e92-b727d1a952c7-must-gather-output\") pod \"must-gather-4hdfc\" (UID: \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\") " pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:46:26 crc kubenswrapper[4718]: I1206 04:46:26.986732 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hgpvg/must-gather-4hdfc"] Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.077350 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdjr4\" (UniqueName: \"kubernetes.io/projected/7ef1b93f-22cf-446a-8e92-b727d1a952c7-kube-api-access-kdjr4\") pod \"must-gather-4hdfc\" (UID: \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\") " pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.077428 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ef1b93f-22cf-446a-8e92-b727d1a952c7-must-gather-output\") pod \"must-gather-4hdfc\" (UID: \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\") " pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.077885 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ef1b93f-22cf-446a-8e92-b727d1a952c7-must-gather-output\") pod \"must-gather-4hdfc\" (UID: \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\") " pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.178725 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdjr4\" (UniqueName: \"kubernetes.io/projected/7ef1b93f-22cf-446a-8e92-b727d1a952c7-kube-api-access-kdjr4\") pod \"must-gather-4hdfc\" (UID: \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\") " pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.199682 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdjr4\" (UniqueName: \"kubernetes.io/projected/7ef1b93f-22cf-446a-8e92-b727d1a952c7-kube-api-access-kdjr4\") pod \"must-gather-4hdfc\" (UID: \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\") " pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.289405 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.470456 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hgpvg/must-gather-4hdfc"] Dec 06 04:46:27 crc kubenswrapper[4718]: W1206 04:46:27.478445 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ef1b93f_22cf_446a_8e92_b727d1a952c7.slice/crio-8d59f4c58b6be5934b380ca6454b91530e1817a0bd14c65a43b3f8c46e61b158 WatchSource:0}: Error finding container 8d59f4c58b6be5934b380ca6454b91530e1817a0bd14c65a43b3f8c46e61b158: Status 404 returned error can't find the container with id 8d59f4c58b6be5934b380ca6454b91530e1817a0bd14c65a43b3f8c46e61b158 Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.687327 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" event={"ID":"7ef1b93f-22cf-446a-8e92-b727d1a952c7","Type":"ContainerStarted","Data":"8d59f4c58b6be5934b380ca6454b91530e1817a0bd14c65a43b3f8c46e61b158"} Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.877347 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:46:27 crc kubenswrapper[4718]: I1206 04:46:27.877399 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:46:28 crc kubenswrapper[4718]: I1206 04:46:28.693153 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" event={"ID":"7ef1b93f-22cf-446a-8e92-b727d1a952c7","Type":"ContainerStarted","Data":"1cd8f6c731d7dd39b8f26aee14ef4f5e7d360d58025d18fde6c778c8d86e2876"} Dec 06 04:46:28 crc kubenswrapper[4718]: I1206 04:46:28.693192 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" event={"ID":"7ef1b93f-22cf-446a-8e92-b727d1a952c7","Type":"ContainerStarted","Data":"05b29b32e76bd36a9bcebf6ca229d99d29e170baf065bdc834464fd861adf667"} Dec 06 04:46:28 crc kubenswrapper[4718]: I1206 04:46:28.709495 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" podStartSLOduration=2.70947835 podStartE2EDuration="2.70947835s" podCreationTimestamp="2025-12-06 04:46:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 04:46:28.707370836 +0000 UTC m=+2377.713075997" watchObservedRunningTime="2025-12-06 04:46:28.70947835 +0000 UTC m=+2377.715183511" Dec 06 04:46:53 crc kubenswrapper[4718]: I1206 04:46:53.885400 4718 scope.go:117] "RemoveContainer" containerID="5200401f5f35796459f3b9f14f820430c7d3ee6cb352636a4f282c2ef4bba2a4" Dec 06 04:46:57 crc kubenswrapper[4718]: I1206 04:46:57.876794 4718 patch_prober.go:28] interesting pod/machine-config-daemon-pr7fz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 04:46:57 crc kubenswrapper[4718]: I1206 04:46:57.878286 4718 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 04:46:57 crc kubenswrapper[4718]: I1206 04:46:57.878395 4718 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" Dec 06 04:46:57 crc kubenswrapper[4718]: I1206 04:46:57.879030 4718 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037"} pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 04:46:57 crc kubenswrapper[4718]: I1206 04:46:57.879425 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerName="machine-config-daemon" containerID="cri-o://d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" gracePeriod=600 Dec 06 04:46:58 crc kubenswrapper[4718]: E1206 04:46:58.000167 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:46:58 crc kubenswrapper[4718]: I1206 04:46:58.884692 4718 generic.go:334] "Generic (PLEG): container finished" podID="87a3c869-d0a2-46cd-ac46-10022d92c7af" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" exitCode=0 Dec 06 04:46:58 crc kubenswrapper[4718]: I1206 04:46:58.884749 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerDied","Data":"d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037"} Dec 06 04:46:58 crc kubenswrapper[4718]: I1206 04:46:58.884793 4718 scope.go:117] "RemoveContainer" containerID="ae94ee04e0f0b6f53a2fb12aac1d020fc0dc1260d9f11235b84486c47305e1c7" Dec 06 04:46:58 crc kubenswrapper[4718]: I1206 04:46:58.885307 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:46:58 crc kubenswrapper[4718]: E1206 04:46:58.885520 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:47:08 crc kubenswrapper[4718]: I1206 04:47:08.203261 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-g2qsr_96226f3a-94e6-4895-a9fc-662f07f8cdf8/control-plane-machine-set-operator/0.log" Dec 06 04:47:08 crc kubenswrapper[4718]: I1206 04:47:08.396563 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6jmn4_5d02801a-a515-4f1f-99a5-90863caec537/kube-rbac-proxy/0.log" Dec 06 04:47:08 crc kubenswrapper[4718]: I1206 04:47:08.421602 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6jmn4_5d02801a-a515-4f1f-99a5-90863caec537/machine-api-operator/0.log" Dec 06 04:47:10 crc kubenswrapper[4718]: I1206 04:47:10.328542 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:47:10 crc kubenswrapper[4718]: E1206 04:47:10.329180 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:47:23 crc kubenswrapper[4718]: I1206 04:47:23.328550 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:47:23 crc kubenswrapper[4718]: E1206 04:47:23.330211 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:47:23 crc kubenswrapper[4718]: I1206 04:47:23.733750 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-r6chz_b00ca948-0cd8-43bc-8b38-43f445044e51/controller/0.log" Dec 06 04:47:23 crc kubenswrapper[4718]: I1206 04:47:23.756168 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-r6chz_b00ca948-0cd8-43bc-8b38-43f445044e51/kube-rbac-proxy/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.034416 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-frr-files/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.215091 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-metrics/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.241587 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-reloader/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.255516 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-frr-files/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.270441 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-reloader/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.410311 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-frr-files/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.410454 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-metrics/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.454498 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-reloader/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.493531 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-metrics/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.669459 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-frr-files/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.692790 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-reloader/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.696860 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/cp-metrics/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.704756 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/controller/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.868459 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/frr-metrics/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.884324 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/kube-rbac-proxy/0.log" Dec 06 04:47:24 crc kubenswrapper[4718]: I1206 04:47:24.934018 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/kube-rbac-proxy-frr/0.log" Dec 06 04:47:25 crc kubenswrapper[4718]: I1206 04:47:25.126147 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/reloader/0.log" Dec 06 04:47:25 crc kubenswrapper[4718]: I1206 04:47:25.154811 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-7swws_0b00eded-0715-4a41-89e9-3f758847a16a/frr-k8s-webhook-server/0.log" Dec 06 04:47:25 crc kubenswrapper[4718]: I1206 04:47:25.240760 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gp6x6_e878e223-5476-4e98-ae10-2db9708f0336/frr/0.log" Dec 06 04:47:25 crc kubenswrapper[4718]: I1206 04:47:25.329369 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-57d558fdb8-rqqm5_939e915d-3ad4-4a15-bb70-01106cd5cb96/manager/0.log" Dec 06 04:47:25 crc kubenswrapper[4718]: I1206 04:47:25.405921 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5c49888cd9-w2wwt_e5e44d0d-be18-4944-ae68-8758042af15e/webhook-server/0.log" Dec 06 04:47:25 crc kubenswrapper[4718]: I1206 04:47:25.504849 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9wqkt_1bc1c47e-1be5-4a69-96cf-fa5c59004f8b/kube-rbac-proxy/0.log" Dec 06 04:47:25 crc kubenswrapper[4718]: I1206 04:47:25.611271 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9wqkt_1bc1c47e-1be5-4a69-96cf-fa5c59004f8b/speaker/0.log" Dec 06 04:47:35 crc kubenswrapper[4718]: I1206 04:47:35.328202 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:47:35 crc kubenswrapper[4718]: E1206 04:47:35.328939 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:47:46 crc kubenswrapper[4718]: I1206 04:47:46.327655 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:47:46 crc kubenswrapper[4718]: E1206 04:47:46.328380 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:47:48 crc kubenswrapper[4718]: I1206 04:47:48.476823 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/util/0.log" Dec 06 04:47:48 crc kubenswrapper[4718]: I1206 04:47:48.619410 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/util/0.log" Dec 06 04:47:48 crc kubenswrapper[4718]: I1206 04:47:48.632307 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/pull/0.log" Dec 06 04:47:48 crc kubenswrapper[4718]: I1206 04:47:48.679974 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/pull/0.log" Dec 06 04:47:48 crc kubenswrapper[4718]: I1206 04:47:48.802186 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/extract/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.080998 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/pull/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.095893 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8392fd7_f08e6b54-9291-41eb-9b5c-4265e8e25eb9/util/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.133188 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-utilities/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.247799 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-content/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.255417 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-content/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.273475 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-utilities/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.419594 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-utilities/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.425504 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/extract-content/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.626446 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-utilities/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.799043 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-utilities/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.807341 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-content/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.846068 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-content/0.log" Dec 06 04:47:49 crc kubenswrapper[4718]: I1206 04:47:49.882770 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-clxct_1c27751f-4adc-45e6-8e0f-1a4ead35f894/registry-server/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.017403 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-content/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.017435 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/extract-utilities/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.216706 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-42x7c_49e186cc-90d7-4d3e-973b-1727f13a1c6d/marketplace-operator/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.275409 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-utilities/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.447673 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-content/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.474091 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-utilities/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.502117 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2sbp7_fb623567-41f3-4c02-837a-40255117f3e8/registry-server/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.547871 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-content/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.648503 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-content/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.649734 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/extract-utilities/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.778354 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bp8fr_15cc3c85-bfb7-4f19-9498-840989a76366/registry-server/0.log" Dec 06 04:47:50 crc kubenswrapper[4718]: I1206 04:47:50.864045 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-utilities/0.log" Dec 06 04:47:51 crc kubenswrapper[4718]: I1206 04:47:51.029792 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-utilities/0.log" Dec 06 04:47:51 crc kubenswrapper[4718]: I1206 04:47:51.032180 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-content/0.log" Dec 06 04:47:51 crc kubenswrapper[4718]: I1206 04:47:51.047632 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-content/0.log" Dec 06 04:47:51 crc kubenswrapper[4718]: I1206 04:47:51.211179 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-utilities/0.log" Dec 06 04:47:51 crc kubenswrapper[4718]: I1206 04:47:51.229296 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/extract-content/0.log" Dec 06 04:47:51 crc kubenswrapper[4718]: I1206 04:47:51.583743 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j784j_04067a51-e169-4754-b847-e26e29ec8213/registry-server/0.log" Dec 06 04:47:53 crc kubenswrapper[4718]: I1206 04:47:53.933505 4718 scope.go:117] "RemoveContainer" containerID="97592df6292574768ce9ee66b00fe813b74502ddb1690548acb60dfe9a272dc4" Dec 06 04:47:53 crc kubenswrapper[4718]: I1206 04:47:53.950683 4718 scope.go:117] "RemoveContainer" containerID="6f18368bc59dc3f6b2f16d03f381bd231417914db48d6323c57f7258d3d9a5dc" Dec 06 04:47:59 crc kubenswrapper[4718]: I1206 04:47:59.328605 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:47:59 crc kubenswrapper[4718]: E1206 04:47:59.329498 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:48:14 crc kubenswrapper[4718]: I1206 04:48:14.328173 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:48:14 crc kubenswrapper[4718]: E1206 04:48:14.329021 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:48:28 crc kubenswrapper[4718]: I1206 04:48:28.329941 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:48:28 crc kubenswrapper[4718]: E1206 04:48:28.330988 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:48:40 crc kubenswrapper[4718]: I1206 04:48:40.328922 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:48:40 crc kubenswrapper[4718]: E1206 04:48:40.330072 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:48:52 crc kubenswrapper[4718]: I1206 04:48:52.327900 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:48:52 crc kubenswrapper[4718]: E1206 04:48:52.328658 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:48:55 crc kubenswrapper[4718]: I1206 04:48:55.593149 4718 generic.go:334] "Generic (PLEG): container finished" podID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" containerID="05b29b32e76bd36a9bcebf6ca229d99d29e170baf065bdc834464fd861adf667" exitCode=0 Dec 06 04:48:55 crc kubenswrapper[4718]: I1206 04:48:55.593220 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" event={"ID":"7ef1b93f-22cf-446a-8e92-b727d1a952c7","Type":"ContainerDied","Data":"05b29b32e76bd36a9bcebf6ca229d99d29e170baf065bdc834464fd861adf667"} Dec 06 04:48:55 crc kubenswrapper[4718]: I1206 04:48:55.594315 4718 scope.go:117] "RemoveContainer" containerID="05b29b32e76bd36a9bcebf6ca229d99d29e170baf065bdc834464fd861adf667" Dec 06 04:48:55 crc kubenswrapper[4718]: I1206 04:48:55.637977 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hgpvg_must-gather-4hdfc_7ef1b93f-22cf-446a-8e92-b727d1a952c7/gather/0.log" Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.445706 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hgpvg/must-gather-4hdfc"] Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.446580 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" podUID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" containerName="copy" containerID="cri-o://1cd8f6c731d7dd39b8f26aee14ef4f5e7d360d58025d18fde6c778c8d86e2876" gracePeriod=2 Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.449374 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hgpvg/must-gather-4hdfc"] Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.664343 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hgpvg_must-gather-4hdfc_7ef1b93f-22cf-446a-8e92-b727d1a952c7/copy/0.log" Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.666045 4718 generic.go:334] "Generic (PLEG): container finished" podID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" containerID="1cd8f6c731d7dd39b8f26aee14ef4f5e7d360d58025d18fde6c778c8d86e2876" exitCode=143 Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.788711 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hgpvg_must-gather-4hdfc_7ef1b93f-22cf-446a-8e92-b727d1a952c7/copy/0.log" Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.789169 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.919123 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdjr4\" (UniqueName: \"kubernetes.io/projected/7ef1b93f-22cf-446a-8e92-b727d1a952c7-kube-api-access-kdjr4\") pod \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\" (UID: \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\") " Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.919180 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ef1b93f-22cf-446a-8e92-b727d1a952c7-must-gather-output\") pod \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\" (UID: \"7ef1b93f-22cf-446a-8e92-b727d1a952c7\") " Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.928869 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ef1b93f-22cf-446a-8e92-b727d1a952c7-kube-api-access-kdjr4" (OuterVolumeSpecName: "kube-api-access-kdjr4") pod "7ef1b93f-22cf-446a-8e92-b727d1a952c7" (UID: "7ef1b93f-22cf-446a-8e92-b727d1a952c7"). InnerVolumeSpecName "kube-api-access-kdjr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:49:04 crc kubenswrapper[4718]: I1206 04:49:04.981991 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ef1b93f-22cf-446a-8e92-b727d1a952c7-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "7ef1b93f-22cf-446a-8e92-b727d1a952c7" (UID: "7ef1b93f-22cf-446a-8e92-b727d1a952c7"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:49:05 crc kubenswrapper[4718]: I1206 04:49:05.020864 4718 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ef1b93f-22cf-446a-8e92-b727d1a952c7-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 04:49:05 crc kubenswrapper[4718]: I1206 04:49:05.020898 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdjr4\" (UniqueName: \"kubernetes.io/projected/7ef1b93f-22cf-446a-8e92-b727d1a952c7-kube-api-access-kdjr4\") on node \"crc\" DevicePath \"\"" Dec 06 04:49:05 crc kubenswrapper[4718]: I1206 04:49:05.334610 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" path="/var/lib/kubelet/pods/7ef1b93f-22cf-446a-8e92-b727d1a952c7/volumes" Dec 06 04:49:05 crc kubenswrapper[4718]: I1206 04:49:05.672929 4718 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hgpvg_must-gather-4hdfc_7ef1b93f-22cf-446a-8e92-b727d1a952c7/copy/0.log" Dec 06 04:49:05 crc kubenswrapper[4718]: I1206 04:49:05.673341 4718 scope.go:117] "RemoveContainer" containerID="1cd8f6c731d7dd39b8f26aee14ef4f5e7d360d58025d18fde6c778c8d86e2876" Dec 06 04:49:05 crc kubenswrapper[4718]: I1206 04:49:05.673441 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hgpvg/must-gather-4hdfc" Dec 06 04:49:05 crc kubenswrapper[4718]: I1206 04:49:05.697629 4718 scope.go:117] "RemoveContainer" containerID="05b29b32e76bd36a9bcebf6ca229d99d29e170baf065bdc834464fd861adf667" Dec 06 04:49:06 crc kubenswrapper[4718]: I1206 04:49:06.328224 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:49:06 crc kubenswrapper[4718]: E1206 04:49:06.328626 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:49:18 crc kubenswrapper[4718]: I1206 04:49:18.328061 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:49:18 crc kubenswrapper[4718]: E1206 04:49:18.329026 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:49:31 crc kubenswrapper[4718]: I1206 04:49:31.333401 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:49:31 crc kubenswrapper[4718]: E1206 04:49:31.334345 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:49:45 crc kubenswrapper[4718]: I1206 04:49:45.328167 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:49:45 crc kubenswrapper[4718]: E1206 04:49:45.328947 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:49:59 crc kubenswrapper[4718]: I1206 04:49:59.327652 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:49:59 crc kubenswrapper[4718]: E1206 04:49:59.328398 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:50:13 crc kubenswrapper[4718]: I1206 04:50:13.328932 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:50:13 crc kubenswrapper[4718]: E1206 04:50:13.330019 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:50:24 crc kubenswrapper[4718]: I1206 04:50:24.328576 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:50:24 crc kubenswrapper[4718]: E1206 04:50:24.329456 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:50:38 crc kubenswrapper[4718]: I1206 04:50:38.329191 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:50:38 crc kubenswrapper[4718]: E1206 04:50:38.330337 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:50:53 crc kubenswrapper[4718]: I1206 04:50:53.329128 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:50:53 crc kubenswrapper[4718]: E1206 04:50:53.330135 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:51:05 crc kubenswrapper[4718]: I1206 04:51:05.328528 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:51:05 crc kubenswrapper[4718]: E1206 04:51:05.329403 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.085202 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t4xps"] Dec 06 04:51:15 crc kubenswrapper[4718]: E1206 04:51:15.085790 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" containerName="gather" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.085800 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" containerName="gather" Dec 06 04:51:15 crc kubenswrapper[4718]: E1206 04:51:15.085809 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" containerName="copy" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.085815 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" containerName="copy" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.085910 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" containerName="copy" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.085927 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ef1b93f-22cf-446a-8e92-b727d1a952c7" containerName="gather" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.086816 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.106418 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t4xps"] Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.133536 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-utilities\") pod \"redhat-operators-t4xps\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.133592 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-catalog-content\") pod \"redhat-operators-t4xps\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.133616 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpl87\" (UniqueName: \"kubernetes.io/projected/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-kube-api-access-xpl87\") pod \"redhat-operators-t4xps\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.235340 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-utilities\") pod \"redhat-operators-t4xps\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.235404 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-catalog-content\") pod \"redhat-operators-t4xps\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.235449 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpl87\" (UniqueName: \"kubernetes.io/projected/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-kube-api-access-xpl87\") pod \"redhat-operators-t4xps\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.235912 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-utilities\") pod \"redhat-operators-t4xps\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.236059 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-catalog-content\") pod \"redhat-operators-t4xps\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.253460 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpl87\" (UniqueName: \"kubernetes.io/projected/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-kube-api-access-xpl87\") pod \"redhat-operators-t4xps\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.401766 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:15 crc kubenswrapper[4718]: I1206 04:51:15.614003 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t4xps"] Dec 06 04:51:16 crc kubenswrapper[4718]: I1206 04:51:16.554895 4718 generic.go:334] "Generic (PLEG): container finished" podID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerID="e42ad3562744218764d1a883a786ed61f5975ed9fda7251be5847651a98a945c" exitCode=0 Dec 06 04:51:16 crc kubenswrapper[4718]: I1206 04:51:16.554940 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4xps" event={"ID":"d84ccdef-3b35-4d07-9e51-c80b19c70e2c","Type":"ContainerDied","Data":"e42ad3562744218764d1a883a786ed61f5975ed9fda7251be5847651a98a945c"} Dec 06 04:51:16 crc kubenswrapper[4718]: I1206 04:51:16.554968 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4xps" event={"ID":"d84ccdef-3b35-4d07-9e51-c80b19c70e2c","Type":"ContainerStarted","Data":"cab3eb9dd938fe5dce1dd66ebac9784c18d57d4d611b40ffd9d0373c7a4c85c0"} Dec 06 04:51:16 crc kubenswrapper[4718]: I1206 04:51:16.556873 4718 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 04:51:17 crc kubenswrapper[4718]: I1206 04:51:17.561800 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4xps" event={"ID":"d84ccdef-3b35-4d07-9e51-c80b19c70e2c","Type":"ContainerStarted","Data":"31792d7c58bb32c5bb36f1e4166610f519dc672ecae78401692c198aabc325a4"} Dec 06 04:51:18 crc kubenswrapper[4718]: I1206 04:51:18.570958 4718 generic.go:334] "Generic (PLEG): container finished" podID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerID="31792d7c58bb32c5bb36f1e4166610f519dc672ecae78401692c198aabc325a4" exitCode=0 Dec 06 04:51:18 crc kubenswrapper[4718]: I1206 04:51:18.571011 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4xps" event={"ID":"d84ccdef-3b35-4d07-9e51-c80b19c70e2c","Type":"ContainerDied","Data":"31792d7c58bb32c5bb36f1e4166610f519dc672ecae78401692c198aabc325a4"} Dec 06 04:51:19 crc kubenswrapper[4718]: I1206 04:51:19.328009 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:51:19 crc kubenswrapper[4718]: E1206 04:51:19.328565 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:51:19 crc kubenswrapper[4718]: I1206 04:51:19.580953 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4xps" event={"ID":"d84ccdef-3b35-4d07-9e51-c80b19c70e2c","Type":"ContainerStarted","Data":"0e1139b1a28f6bd2d14587e723cc6d1fc13a493b5ce93eda44076099167ca093"} Dec 06 04:51:19 crc kubenswrapper[4718]: I1206 04:51:19.609754 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t4xps" podStartSLOduration=2.183260592 podStartE2EDuration="4.609734659s" podCreationTimestamp="2025-12-06 04:51:15 +0000 UTC" firstStartedPulling="2025-12-06 04:51:16.55666109 +0000 UTC m=+2665.562366251" lastFinishedPulling="2025-12-06 04:51:18.983135127 +0000 UTC m=+2667.988840318" observedRunningTime="2025-12-06 04:51:19.60865018 +0000 UTC m=+2668.614355371" watchObservedRunningTime="2025-12-06 04:51:19.609734659 +0000 UTC m=+2668.615439820" Dec 06 04:51:25 crc kubenswrapper[4718]: I1206 04:51:25.402097 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:25 crc kubenswrapper[4718]: I1206 04:51:25.402756 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:25 crc kubenswrapper[4718]: I1206 04:51:25.479862 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:25 crc kubenswrapper[4718]: I1206 04:51:25.652111 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:25 crc kubenswrapper[4718]: I1206 04:51:25.718490 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t4xps"] Dec 06 04:51:27 crc kubenswrapper[4718]: I1206 04:51:27.626430 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t4xps" podUID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerName="registry-server" containerID="cri-o://0e1139b1a28f6bd2d14587e723cc6d1fc13a493b5ce93eda44076099167ca093" gracePeriod=2 Dec 06 04:51:29 crc kubenswrapper[4718]: I1206 04:51:29.641355 4718 generic.go:334] "Generic (PLEG): container finished" podID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerID="0e1139b1a28f6bd2d14587e723cc6d1fc13a493b5ce93eda44076099167ca093" exitCode=0 Dec 06 04:51:29 crc kubenswrapper[4718]: I1206 04:51:29.641793 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4xps" event={"ID":"d84ccdef-3b35-4d07-9e51-c80b19c70e2c","Type":"ContainerDied","Data":"0e1139b1a28f6bd2d14587e723cc6d1fc13a493b5ce93eda44076099167ca093"} Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.123692 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.325813 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpl87\" (UniqueName: \"kubernetes.io/projected/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-kube-api-access-xpl87\") pod \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.326261 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-utilities\") pod \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.326295 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-catalog-content\") pod \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\" (UID: \"d84ccdef-3b35-4d07-9e51-c80b19c70e2c\") " Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.328376 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-utilities" (OuterVolumeSpecName: "utilities") pod "d84ccdef-3b35-4d07-9e51-c80b19c70e2c" (UID: "d84ccdef-3b35-4d07-9e51-c80b19c70e2c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.329280 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:51:30 crc kubenswrapper[4718]: E1206 04:51:30.329808 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.334202 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-kube-api-access-xpl87" (OuterVolumeSpecName: "kube-api-access-xpl87") pod "d84ccdef-3b35-4d07-9e51-c80b19c70e2c" (UID: "d84ccdef-3b35-4d07-9e51-c80b19c70e2c"). InnerVolumeSpecName "kube-api-access-xpl87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.428200 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.428257 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpl87\" (UniqueName: \"kubernetes.io/projected/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-kube-api-access-xpl87\") on node \"crc\" DevicePath \"\"" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.493242 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d84ccdef-3b35-4d07-9e51-c80b19c70e2c" (UID: "d84ccdef-3b35-4d07-9e51-c80b19c70e2c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.529394 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d84ccdef-3b35-4d07-9e51-c80b19c70e2c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.650683 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4xps" event={"ID":"d84ccdef-3b35-4d07-9e51-c80b19c70e2c","Type":"ContainerDied","Data":"cab3eb9dd938fe5dce1dd66ebac9784c18d57d4d611b40ffd9d0373c7a4c85c0"} Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.652093 4718 scope.go:117] "RemoveContainer" containerID="0e1139b1a28f6bd2d14587e723cc6d1fc13a493b5ce93eda44076099167ca093" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.650747 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4xps" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.673251 4718 scope.go:117] "RemoveContainer" containerID="31792d7c58bb32c5bb36f1e4166610f519dc672ecae78401692c198aabc325a4" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.699748 4718 scope.go:117] "RemoveContainer" containerID="e42ad3562744218764d1a883a786ed61f5975ed9fda7251be5847651a98a945c" Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.713998 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t4xps"] Dec 06 04:51:30 crc kubenswrapper[4718]: I1206 04:51:30.721949 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t4xps"] Dec 06 04:51:31 crc kubenswrapper[4718]: I1206 04:51:31.354749 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" path="/var/lib/kubelet/pods/d84ccdef-3b35-4d07-9e51-c80b19c70e2c/volumes" Dec 06 04:51:45 crc kubenswrapper[4718]: I1206 04:51:45.329252 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:51:45 crc kubenswrapper[4718]: E1206 04:51:45.329859 4718 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pr7fz_openshift-machine-config-operator(87a3c869-d0a2-46cd-ac46-10022d92c7af)\"" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" podUID="87a3c869-d0a2-46cd-ac46-10022d92c7af" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.126579 4718 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r4wkd"] Dec 06 04:51:55 crc kubenswrapper[4718]: E1206 04:51:55.127673 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerName="extract-content" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.127698 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerName="extract-content" Dec 06 04:51:55 crc kubenswrapper[4718]: E1206 04:51:55.127715 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerName="registry-server" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.127727 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerName="registry-server" Dec 06 04:51:55 crc kubenswrapper[4718]: E1206 04:51:55.127763 4718 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerName="extract-utilities" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.127779 4718 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerName="extract-utilities" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.127959 4718 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84ccdef-3b35-4d07-9e51-c80b19c70e2c" containerName="registry-server" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.129511 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.146118 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r4wkd"] Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.308638 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-catalog-content\") pod \"certified-operators-r4wkd\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.308704 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slfzd\" (UniqueName: \"kubernetes.io/projected/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-kube-api-access-slfzd\") pod \"certified-operators-r4wkd\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.308741 4718 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-utilities\") pod \"certified-operators-r4wkd\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.410458 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-catalog-content\") pod \"certified-operators-r4wkd\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.410518 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slfzd\" (UniqueName: \"kubernetes.io/projected/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-kube-api-access-slfzd\") pod \"certified-operators-r4wkd\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.410562 4718 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-utilities\") pod \"certified-operators-r4wkd\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.411068 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-catalog-content\") pod \"certified-operators-r4wkd\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.411076 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-utilities\") pod \"certified-operators-r4wkd\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.434095 4718 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slfzd\" (UniqueName: \"kubernetes.io/projected/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-kube-api-access-slfzd\") pod \"certified-operators-r4wkd\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.472656 4718 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:51:55 crc kubenswrapper[4718]: I1206 04:51:55.919998 4718 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r4wkd"] Dec 06 04:51:55 crc kubenswrapper[4718]: W1206 04:51:55.933514 4718 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd40ed445_dc89_4a17_9f86_7cad1eb3af1a.slice/crio-a4d06d613d9d0dcf3d127396c26b4529ff6518fa97b10c0c8d3364ecedcb2534 WatchSource:0}: Error finding container a4d06d613d9d0dcf3d127396c26b4529ff6518fa97b10c0c8d3364ecedcb2534: Status 404 returned error can't find the container with id a4d06d613d9d0dcf3d127396c26b4529ff6518fa97b10c0c8d3364ecedcb2534 Dec 06 04:51:56 crc kubenswrapper[4718]: I1206 04:51:56.833280 4718 generic.go:334] "Generic (PLEG): container finished" podID="d40ed445-dc89-4a17-9f86-7cad1eb3af1a" containerID="7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45" exitCode=0 Dec 06 04:51:56 crc kubenswrapper[4718]: I1206 04:51:56.837820 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4wkd" event={"ID":"d40ed445-dc89-4a17-9f86-7cad1eb3af1a","Type":"ContainerDied","Data":"7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45"} Dec 06 04:51:56 crc kubenswrapper[4718]: I1206 04:51:56.841945 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4wkd" event={"ID":"d40ed445-dc89-4a17-9f86-7cad1eb3af1a","Type":"ContainerStarted","Data":"a4d06d613d9d0dcf3d127396c26b4529ff6518fa97b10c0c8d3364ecedcb2534"} Dec 06 04:51:58 crc kubenswrapper[4718]: I1206 04:51:58.854975 4718 generic.go:334] "Generic (PLEG): container finished" podID="d40ed445-dc89-4a17-9f86-7cad1eb3af1a" containerID="395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763" exitCode=0 Dec 06 04:51:58 crc kubenswrapper[4718]: I1206 04:51:58.855034 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4wkd" event={"ID":"d40ed445-dc89-4a17-9f86-7cad1eb3af1a","Type":"ContainerDied","Data":"395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763"} Dec 06 04:51:59 crc kubenswrapper[4718]: I1206 04:51:59.863225 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4wkd" event={"ID":"d40ed445-dc89-4a17-9f86-7cad1eb3af1a","Type":"ContainerStarted","Data":"020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57"} Dec 06 04:51:59 crc kubenswrapper[4718]: I1206 04:51:59.878873 4718 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r4wkd" podStartSLOduration=2.463620454 podStartE2EDuration="4.8788533s" podCreationTimestamp="2025-12-06 04:51:55 +0000 UTC" firstStartedPulling="2025-12-06 04:51:56.83509184 +0000 UTC m=+2705.840797001" lastFinishedPulling="2025-12-06 04:51:59.250324696 +0000 UTC m=+2708.256029847" observedRunningTime="2025-12-06 04:51:59.878067798 +0000 UTC m=+2708.883772969" watchObservedRunningTime="2025-12-06 04:51:59.8788533 +0000 UTC m=+2708.884558461" Dec 06 04:52:00 crc kubenswrapper[4718]: I1206 04:52:00.328243 4718 scope.go:117] "RemoveContainer" containerID="d68b9a2cf798dd63dbbc0f84c0a90be0700d987e2a9e77e49845b04d5f37e037" Dec 06 04:52:00 crc kubenswrapper[4718]: I1206 04:52:00.876572 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pr7fz" event={"ID":"87a3c869-d0a2-46cd-ac46-10022d92c7af","Type":"ContainerStarted","Data":"275849c2538ed96067c948b547e9aa436f8ff411e43cdf1d746e8695ccc36715"} Dec 06 04:52:05 crc kubenswrapper[4718]: I1206 04:52:05.473006 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:52:05 crc kubenswrapper[4718]: I1206 04:52:05.474903 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:52:05 crc kubenswrapper[4718]: I1206 04:52:05.536512 4718 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:52:05 crc kubenswrapper[4718]: I1206 04:52:05.969003 4718 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:52:06 crc kubenswrapper[4718]: I1206 04:52:06.022150 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r4wkd"] Dec 06 04:52:07 crc kubenswrapper[4718]: I1206 04:52:07.926615 4718 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r4wkd" podUID="d40ed445-dc89-4a17-9f86-7cad1eb3af1a" containerName="registry-server" containerID="cri-o://020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57" gracePeriod=2 Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.896662 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.911529 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slfzd\" (UniqueName: \"kubernetes.io/projected/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-kube-api-access-slfzd\") pod \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.911620 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-utilities\") pod \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.911712 4718 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-catalog-content\") pod \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\" (UID: \"d40ed445-dc89-4a17-9f86-7cad1eb3af1a\") " Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.916511 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-utilities" (OuterVolumeSpecName: "utilities") pod "d40ed445-dc89-4a17-9f86-7cad1eb3af1a" (UID: "d40ed445-dc89-4a17-9f86-7cad1eb3af1a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.926615 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-kube-api-access-slfzd" (OuterVolumeSpecName: "kube-api-access-slfzd") pod "d40ed445-dc89-4a17-9f86-7cad1eb3af1a" (UID: "d40ed445-dc89-4a17-9f86-7cad1eb3af1a"). InnerVolumeSpecName "kube-api-access-slfzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.936734 4718 generic.go:334] "Generic (PLEG): container finished" podID="d40ed445-dc89-4a17-9f86-7cad1eb3af1a" containerID="020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57" exitCode=0 Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.936771 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4wkd" event={"ID":"d40ed445-dc89-4a17-9f86-7cad1eb3af1a","Type":"ContainerDied","Data":"020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57"} Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.936793 4718 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r4wkd" event={"ID":"d40ed445-dc89-4a17-9f86-7cad1eb3af1a","Type":"ContainerDied","Data":"a4d06d613d9d0dcf3d127396c26b4529ff6518fa97b10c0c8d3364ecedcb2534"} Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.936808 4718 scope.go:117] "RemoveContainer" containerID="020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.936931 4718 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r4wkd" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.958068 4718 scope.go:117] "RemoveContainer" containerID="395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.975284 4718 scope.go:117] "RemoveContainer" containerID="7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.996892 4718 scope.go:117] "RemoveContainer" containerID="020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57" Dec 06 04:52:08 crc kubenswrapper[4718]: E1206 04:52:08.997392 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57\": container with ID starting with 020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57 not found: ID does not exist" containerID="020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.997450 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57"} err="failed to get container status \"020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57\": rpc error: code = NotFound desc = could not find container \"020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57\": container with ID starting with 020e1c108cbd9421e7dfab8b987e381067ee665183bd18316bd4c9bb43356b57 not found: ID does not exist" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.997483 4718 scope.go:117] "RemoveContainer" containerID="395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763" Dec 06 04:52:08 crc kubenswrapper[4718]: E1206 04:52:08.997833 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763\": container with ID starting with 395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763 not found: ID does not exist" containerID="395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.997870 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763"} err="failed to get container status \"395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763\": rpc error: code = NotFound desc = could not find container \"395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763\": container with ID starting with 395c447135256c461da382583196104e45d539de9a262b12da10bb96c6727763 not found: ID does not exist" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.997900 4718 scope.go:117] "RemoveContainer" containerID="7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45" Dec 06 04:52:08 crc kubenswrapper[4718]: E1206 04:52:08.998830 4718 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45\": container with ID starting with 7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45 not found: ID does not exist" containerID="7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45" Dec 06 04:52:08 crc kubenswrapper[4718]: I1206 04:52:08.998884 4718 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45"} err="failed to get container status \"7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45\": rpc error: code = NotFound desc = could not find container \"7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45\": container with ID starting with 7d47f7854a697a6ab85e446fd5648ca79efbcf7f900acfd01bb7e8ed74bb2a45 not found: ID does not exist" Dec 06 04:52:09 crc kubenswrapper[4718]: I1206 04:52:09.005341 4718 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d40ed445-dc89-4a17-9f86-7cad1eb3af1a" (UID: "d40ed445-dc89-4a17-9f86-7cad1eb3af1a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 04:52:09 crc kubenswrapper[4718]: I1206 04:52:09.014492 4718 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:09 crc kubenswrapper[4718]: I1206 04:52:09.014533 4718 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:09 crc kubenswrapper[4718]: I1206 04:52:09.014572 4718 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slfzd\" (UniqueName: \"kubernetes.io/projected/d40ed445-dc89-4a17-9f86-7cad1eb3af1a-kube-api-access-slfzd\") on node \"crc\" DevicePath \"\"" Dec 06 04:52:09 crc kubenswrapper[4718]: I1206 04:52:09.300602 4718 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r4wkd"] Dec 06 04:52:09 crc kubenswrapper[4718]: I1206 04:52:09.305002 4718 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r4wkd"] Dec 06 04:52:09 crc kubenswrapper[4718]: I1206 04:52:09.338850 4718 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d40ed445-dc89-4a17-9f86-7cad1eb3af1a" path="/var/lib/kubelet/pods/d40ed445-dc89-4a17-9f86-7cad1eb3af1a/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114733241024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114733242017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114725423016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114725423015461 5ustar corecore